var/home/core/zuul-output/0000755000175000017500000000000015134370011014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134375735015506 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000235043115134375576020276 0ustar corecore~qikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD >"mv?_eGbuuțx{w7ݭ7֫~𒆷7̗8zTY\].f}嗷ovϷw_>on3cvX~egQBeH,nWb m/m}*L~AzHev_uαHJ2E$(Ͽ|/+k*z>p R⥑gF)49)(oՈ7_k0m^p9PneQn͂YEeeɹ ^ʙ|ʕ0MۂAraZR}@E1%]˜(O)X(6I;Ff"mcI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|iZ~hal\t2Hgb*t--ߖ|Hp(-J C?>:zR{܃ lM6_OފߍO1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;o_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾgm\Sj#3hEEH*Nf äE@O0~y[쾋t=iYhșC 5ܩa!ǛfGtzz*з 55E9Fa?Zk80ݞN|:AОNo;Ⱦzu\0Ac/T%;m ~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0,d%1}Uhs;H?)M"뛲@.Cs*H _0:P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6,֗&̴+s~x?53!}~Z[F)RH?uvͪ _5l *7h?cF_]CNnW)F5d,0SSNK9ް4:ÒozsB<^+鄌4:B%cXhK I}!5 YM%o<>"ہ)Za@Ι}YJz{ɛr|hxY/O$Zøu32EʉD'MS1}t i:Y`cФIX0$lη˽`!i:ګPSPٔ3@5;ȕ}PkڪH9' |":", 1Ҫ8 %lg&:2JC!Mjܽ#`PJWP4Q2:IGӸۡshN+60#:mufe߿~Y,iǑ wVq*T+ w%fx6 %u̩1hӰc%AYW ZY~a_6_yWf`rVA,f=A}h&V<UKkZ{iqi :íy˧FR1u)X9 f΁U ~5batx|ELU:T'T[G*ݧ ؽZK̡O6rLmȰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|miȄ{g6R/wD_tՄ.F+HP'AE; R jK=! I,($F{ձ Oy 6EK( EF #31J8mN .TTF9㕴5~RxCy,&v3,JE- ZK䚒}0BOnYr猸p$nu̿ܣRF]^% >8ȕU4ˬZcն a.1wXhxDI:;.^m9W_c.478NBx;c4<ニ)H .Pd^cRqG+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o=%*[.\Mz_s߭AF'is^_&uUm$[[-HI4QCZ5!N&D[uiXk&2Lg&Ս7/6v_cd쿽d@eU XyX_2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcBN j>Pm[E[^oHmmU̸nG Ym݃|M$ 6.x5 TKXbXj-P\jА޴y$j`ROA"Yg"ź)\3mf|ܔMUiU|Ym! #frۺ^c`Ӄ5¶]Vu@f`mAeerfbkv8)T[8"r&i:]-`ZmG`ͰI+ 3?<' TԼfZX\tKzFgC7/cؖ0ǜIz*෮9E.X{TMYtt =0AMUk}G9TA,;Tt,"Dxl DfA\w; &WaEN^x6mD[%Z`'#^jT_kPW1mk%?\@R>XCl}b ,8; :.b9m]XaINE`!6uOhUuta^xJ@˭d- T5 $4ذ:[a>֋&"[ }_Çk;q 7_= `!6зd B0C?]lja~ luqzFȏ>1k*DlS6vP9hS  ehW!>$]0 5_'lg}릅h:@61Xv`Wk}ˍVbuZ`ͲH-nA:m%neDdF O>y_:,eXX맻c5ޖSwe݊O4L)698ac#sĢB\PIPfwJQJ;Qxm &87I,[[(=.+>n` *UP0Sp8:>m(Zx ,c|!0=0{ P*<7.Va*H7MQK"<O%MTTtx袥:\kfeMuVMy̥Q\ګ1F#șcq##rI$I.im+}2Q14塘ZRI&{3TUFp:4TƓ5[۲yzz+ 4D.Ճ`!OTnPFp':η4dMF^a{x_z;jXh ׄ?xt.o:`xeBe?:ύq;Cl| E"Hd"H`A0&dY3 ً[fct[FidxMUY.b=eaIZ=}sQVΖj?+c;j FRrI5%N/K;Dk |Cbm7чsSW_8'{翴RY.~XfY/UG6 !ȴ ,!NB ^wKykg(s$sN' 88`wC3TY f.s{hkN.g^s7`zDc D]ͺ?jgw07'㤸z YJ\Hb9Ɖ„2Hi{(2HFE?*w*hy4ޙM^٫wF(p]EwQzr*! 5F XrO7E[!gJ^.a&HߣaaQÝ$џwyz4}0!gP_nIzMsPIȰքbFEűEK . ˇj".BBX9%5{M Q 1yV|[jfLs FN*t*oӐ57\Vlhl'jZp1~#aY뼧iùth'*sI~'x#bv#՗*1֌IN}YN+VqozmyY4Ӳ2S<(j%\IGλ`ʌ=]ח?Ώ2J0BLzv8D<%P\MUfN-0]"DBQlt #;NMBQ&{&x6 ݲc|!t(ڀ  lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td3>@)p$`Xs@7,?ע|tC3.㤣TiHEIǢƅaeGF$ u2`d)/-st{E1kٌS*#¦۵_Vu3ЩpRIDr/TxF8'4sѓ{%w μʕ+8u)-" ><-uˮ◶>waPcPw3``m- } vS¢=j=1 7=&=;JW(7b ?+]m`F3 W((!5F-9]dDqL&RΖd}})7 k11 K ;%v'_3 dG8d t#MTU']h7^)O>?~?_ȿM4ə#a&Xi`O}6a-xm`8@;of,![0-7 4f kUy:M֖Esa./zʕy[/ݩqz2¼&'QxJE{cZ7C:?pM z*"#窾+ HsOt۩%͟A498SwWv|jNQ=-ci,̭9vb$ΗDI gV T V%1T\? &!d2ER=j1R=:E mqŗy}d0!ֽ9lX #QYd.x73/s JH;UTd ʲ)97a{nT!S1t&LNNw(c` sd\YXuxJV:p|,U2h\՟ eI}Wn6ٴP?M/ÓUVULF"k3,K- Gk7տu|1,% "HMa"wB,#iM[yy& !"k CGº,4<wɀs"oJWxUCO̫pAG kBc6y_ =JB)a]bB>*`yCBA6Z D֠Wwٞ%eOi9ukCǺ PQgE]qmr+E^nէ')YcI(4Uy(%ěj }11X+x焓!61Bǭ8b><ON1Е7 R@1KVC|ǽ2VCxK;Uh~2gfu#x~]Zݗe$?#`6 )1 >; +RLKyGaJI=;] Iee9]jxxvxHf.hxw}ߜ4:k_ imv0hhg k`EA$@p=wcMԧ y(e QFp8ܑmw*9COt]﮶Ռr{00ya}N{/1_ }VOdVL(c6$o*yF!K wՁ;X"}A楎 ut{D/(=a`V}}O+\69&U1i{JnK$|,Y %H2;sڳJ3F{yՔEHV|ӝ\Eph؈y1LА18U-;s7D h@5}~2E~=T'+܇q2:=^¿չсxtA%Оy"ؤ$x4|>7ZkYFEU6#뮗;"(ϬC{Ʒ A5xꕹꜱ =k{$*j:1зo ! ؾ!9@9S2 )0ߢ5ڦ=:dMuFΆmΙO]4*&;; ,b`E!9bD):6'ǀwZ.j."e uKN&[MQcw:֭ON8pzʳϚW̩{f."R" (d%Wp\Zt=%R2wË-hh7FA wԻޭhd1Ȼ#x{YPC)j_>)$xRSB֪*^3[ 6NE 2M^ hb$ls^OPCU_L~Tb_,njʽw PL!&Zp1EśI^SSB1[bW M` 7#@Pb."+'[\FV.AD2'I1yGow5[\/iI7YwGh)2tޞܚM{L+RC`aHH^.3=#-"8$tG<&\'XLAf4Y=f9r7'9A8ܑܦd;\?tNGw\^.\hPOΉ:Cf9dTyRMZe g]%}V\M6+ FJ7%cهG/!#Z]ooa&!Do`zw6@}t51D 8 %m_|Ys<;suG-=ۮOF\dY(7/'8V%))IHw=].5km8iysNO}b y9%y˻MMX[ROFgG/'#zKW쳮y ڼ%YQF 9M!yɇf "0)#s6ߕf T'}Bl0ϊeG>'Ǜ}}( Qm4^GfR< 0y Dn@$w,66=T{'f%ZMH>ȻwQd뒐BU=Rb]P Hnh|G?$e>P$'г 3o^.G;MR9.[G nN2>x8ĪxK,`} ~a3קZyg 5E5mAњx!37{F|‚k>/F?rʘZ֣Kf9U@!? cNtNm+C 2Ztԓa ݿV6ycQF9lhf3GD}Qᨻ#UaK $cbgYj筃kݼO~f ^HMKg?:9;x۾U;޵oz+O8者]ͮcR=n\t31'úly(cwuW6w/l|/}s;ۢ~}:M x{j>gX /RheX[EZNO|/!}'}iE`]Eކ{w8//o|?n]9>3yPH& P1dR߀']i6#wwO{. ӬfXiߒ`QB;*"c}nq_ 'UȸB&$wZH JF"EcNM4M`ڔä>tc_ `{{J(hjnʵ&jn+=x?]b"wW>;ٍ;+CU ?dWaG!j-'w0s"r6z߹mzRګ$ܱ&Qj^TcLO ^vw'K vP4pɓz^#+CT4l|8 H|pevA;)Y ܣRp!|v"|{/MC4L;M"cȚ7K/FXKsN?j"Xg!`MѐcSh4{@A'aՖQ_~3ҧ&rMR X cb%LXJP)@V$΃"J?n?lE1Nn2 +qVhvrR"f/ZKb{(!dH/d>, P.Df~=PM<')z#;T>nn6 <-q5BHjL5 _ [ʘEӘ8W1/W=O-z>eU3imEhS .bpD}}Kj#Z"DUIowU>IC S%MJ-`Bz^RŬ4"iLoS* P (y.M饸ĕͻ~t2>HoաԮ ?/\dٺ3Mh>屢Q7 jwI>? o~#>n-i3,1ֿk0$OowAiEIv8FH|6 IqN7ͬ5kݰ-Қ”W >Nȶb9 8yݪkIf+;8>V#ہll;/ؽnA(ȱ`Aj1U)XOS k9mUxBa"'4T[Jl /k9rCAj_UiǘP:4F%_0E5IE'rX-|_W8]ʐ/=ӥjhO鞰%?| :S PxbFOor^3>Uf1w;mCja:-1_kٙ5VbZ˙#G6 `q+MPU~ll!.We$9 -D087?a@PV5Bc}jc2FIեmbTbc[O`uxp  rT!|ik3ȱ WW/ ,@a#LA4.;^XۋX^ٝ:^Izq) ٽ<ƎDnv9Bc5Lt3#i RAٽ%| cbpcTfp9 6L/>_x 'ۙz7p~);qU9GLT!o 6]c_:VlnEUdn̶UˇKU ;V`FUݵޙEO[)ܾCy*8ª[cեjGx8&? ՃJȖ9![~[?& "sKT62i sLq9z@JpM->?# n@6# C¾窜̀'n 3u0_k@BS %fee}i^]>¤*l :\gm&iEzyIAtRu~ptոʧߋj\Z?xZTE.1_V߫{]MLmc 6&|)uSy}P~澼2wRw߯Uqw}>g8? t[|W*8n?[>;~mcog]'"U;gm1?Z֒Vv`!2XܬZR`1eL~dپ&+<в~Ɯ,v~,b1)up)J \yS qY>W RzB>+q&[ڲ&<C ka<A+U 3G`vqmJc,Ih: ݪkNwN2taGϰZg ϫ ۷> f+<|d=`h$fZj(ЊZzc \ d5;hLj'KQ %8Gk`N;SlRh)-؈Vnx1Bd="kԦ:,EVuwN ًs9d#Ϸ$*u1>I#bX9~W !X&H2lkVykZt<б vjf/AfO vRPػ.6<-'"vdv .z{I>|!4Ăw< [P{]"cr1殲)6HebLϧB|l_c6.v`V ޾W\N+ď1tq11W vzZ U'=҅crZ:T#\l_:ď);kX!LHmQ (6c9@e|u$4a?"] `W+m𢛲`Rs _I@U8j{.g߷1kY`7;fܑMYٚ/2Jh)ݸVl0/eEU"Ik dRu˜ⶄU:ǖ xs%P f\ w۶q|]Riĥ zlm/bcK3;=,w}RqT vvFI O]&5mKMf#pBTkvyi*,cem:yW|1uX CWL;oG^R X5.aRoߦ[_VscxIޮsPiM9VkίE|s`poSfb|RK> 1GfTq3J&J@Q v&Pߚ񾅶szR.怕#H% }sgvO"4Mی+Ĥ'cp60ǚc6:$[|a.f0G*0-)ZyT13`tTnm* "1LӰ7&jDkM~# (C>ϭQ3Gߤ5UN;?P%ٱm -{2k 8Vbv"ŏݙmn&bX[cO;R`RAd+w!uy9H] bX9^wCο -~d+pLUgRy2Xo3v!ؒRKfs%(1Lhrٵ L.۪] sY4Ew[xٙ,19V"R1I>aJ` UL'5m1Ԥ:t6I >jz( @kUc\ǔ`րj&*ߗEЍ0.U#X) b0G'+1&1 7ɜc:xF 2侯}>Ir~Z#+A<й l~R9]B2o2=>Z6=V*Gg"*5uP|@cK!0 hRh):gnp6L!6^ʁX#nlìRXY\Jk`UZӏL_?C.U0E [%W%~l x٘ ܎y<dY״͎jiTI!jp .Pٱh s@/zIaZfnt^ևeb̯[qM!:ɳCBcd7u-jqX(;(6I#z-2/S zp!m?e8 "(\N+:Y EaSL]<^(]В|Dž8"oRs?]\McZϕ!1hKs`h0OG!L-wCln2&0Y'0=/T4G7! H/ͺ|@bX)+{{T/seϙK$s`h nO/cE(Tcr!:8UL | 8 !:V^S7;G{u%[^g#pnX#&ߓTo_=: sE0ڊSq:+Ekٕ|dvv?jjy6X î/ML-yz,ZlQ^oAnv -)xǺč--qcl @Վ*V߳`Ƅ(Nc}*ۈGmU`tiz5Ӄ~ &$yx `iKA"*1 [cplm_ugTvFW/ɚپi?~?jVXioRA]*pW㺽~2w_|/kplomunJswkQ=\UrmT6S[~k}}#7עz1}qwo確.__W70Z==ݫ˜WOoRy*4݀wn&8֨OreXMlk)=VzpO֠24h 1hi D{q>~%̈v^nBi~MefZN >:/?/Ac 1M/ e9d ǙB?K_z>r%g3JQK-͗R>Kk.Oq,*kTa9xW!8+J&[V =͋Az`sJ|LV/rʑ=IhM4fGQIc(<ǹJi lc][oȒ+_%H6 d'N8hMkͰIٚ[MRd[nY$Itbu] vrJZP'$i|dG UarVF,$| 1bu{Glfa1+cYIe;a6(`nfGd<3 ®+4|<)VR %jբLyHI \ܔ(x%! _l=ݲp; B[qo)4mŚV&X<EQ `V,z5sQAKb#1u c#)yBGՌH,lKb/RoXƣKF*=ݍx(&b<%xrK*{]a/yIu2y9Y\^@M-gKGlMJUjQ]6q %$v-Y{KƣIJ,s(!uU9D{oyfԴ |kzdt [F*xt;yOAQoG-5%@Tv{cdջx| J@9{zw K&dz( َ&_$UOGl }i1.|tKYH}!ҋ㪬v2(4$UH' ^!qKl'gK`9i5 .iSl'VBL{ z/ O:鍃"*zYx^8 Kc~VB/$YZ^o,<dV.̟!S^<[xiERiQxk߯BX^i!fg\D#MЋn gbc<]Ů6(k m"ʺV]#> Q]]_t~~Ȯ:^B8 ce4ҊB9C_4MHn% \1ecJ}gޜ75w DT֕?[P`xO8L"˛eU2eJs080Ogu~} 'oPvW]4NۋjK<,kPc:ЌFK:GzLDfE2Eyj#SMi4d` Jv׍OJBv/ޟzi4>>P7 =J-M :4fr#iX@/8nEkɣGR8ۣ"xZֶǵP\$K#CGSסQq" f_4Hf@|df@!tS]H`7tLAC|g$QhQ'ES,BcpZ5{&E XɵԝNJHx6ߺJ"$W ETk4=U&g84mj)@m/VW,f-2RS7v$œZ(tAhvW}Ń:FrwGlԖ!5l~%.4Kt+Ocl\~ DsZDLsUiAäZhiiJ =m"4ՖM0qjfmdXH8X8'6v±v,< ݋]]MbU<{S+WSuo/W (kֶݻV3 jOTH\VWId.0a yF$f2EZ1i\-UZkRHCFTjI[ΖXhOAuDecHRD8ojtO75IߋpmkCɐCrg3&DUxR#["AmR7U@Pmpa:cU}hsES6w yQm_˨=~>^MpqD,H 4\AژIS1rt"[--si" /*+E $9KUz{ך.Z*֝9XרV-fB`,߷fԏc/|+0.ƾFzo1zV.<݉}\ȴqwh֐eךDxgy;oVSK^U둋R첶jY6 > Kʹ dY MXrtm$IWE7 .tqI"NC=v6bD ]Ւq<#d$~ZJ:"FHr+t޴ Dz?VX&yC8<}ɣKmÅbmU-v쟼ʍuCWjꯪvX@Zʺm D@3U!ϊe`ыUd[FY uQUe=.= A-{8ªa7tABp UXiyi!o#¨ߣ▹6u={ŊXy=x=}z-qnV 3swϒq'4N줽&Hf#zd~IAwmH" `(Bd̮Ufډbi% (9irMuEZUite2ph11P=oYO0YZ]>~2}b:=l0Si25[S *g?i~Ii}43TUtQlQj9@]y8ʪߞ>P-!"XUoʉ"_Vu6Db|M.K PZ7~sя??50qgl56\ G6anu P Tiۙunzl5C*膇;@1Ͷl(fGa$qgA#X :hbL~% A v~P"tïY=^մ KFmf&v(+JK./mn1eTm@?72mNM7: 2v' .D϶Fہh{a+n=3i59@1$0A){$nt!qa,&Z  zg@N/  0Wvo3(;˶v:34? tv%b <s߳~&Lps ;ʶ{}GyDt#-@OT0L%p F @q F(٦ uLN- UgTq:6޷%_x3z;]w4g1 nwR ]`\kuO DŽG_:*<_-[d<%mgw.a24iL9>Ṵz(;Fy] =a$oEsa]qvy'tŜ 53o`agdAW̅]@w[%Gsoʍl tƋ'p:c7yBc\~hRڭ@_ī&I>C2t-qANeYsLR{+ @3èP"1֍7IZ@vwa cuwZ9 JЉ?}IAlZ˯@gpLq=E\YzHV|W`s5 C.31Hdʙ EQ , $Q 052sLs 4Qqkˡu! G[=n̢ "kpq6T`/YKbEjN\jQ#ZFg!5LQ-Ƶ鼚(VQ8e?%+z,SpkT 80)ʋŒ< ,+ĜrGy B!)*>~8[Q2ΦiU1֣a4BR -KpAhqqLS͈F͆59\ʃJ12FXi#7r f΢ia"ݷ0xu9S~\1?`;r *B4 9*Eo=^ oPbjܧpGRP]V*BGiVDxJ/`uU!9d]2 X/>@/皭\VW՗$ xFl(gZC~|yTFLK`sM[Vns<>f4[-':r\gócrHl1ؠA撸 ,>8=IN%N34qU\BJ;op{|y T\KVVO>~<~tܤ!$z1wpڑx'Mㅃ$\'V(ꁲzs,aš" p`8yE-&~zu[6syz73 ''$/nQ2mS!|Vg(=hfWח X$^iq&?e0*|s%]< }Vo`(K]# eFtniLΥ @pæ9xO@?3Q|0K/x6 '欇Yҋ<'~?I'Aj^Фpw5.x &~ x10l\/p}J~;uk8J30v=R~;u r c8i̕UN-U#O 7`\ޑm@u:4t)ꀻQ W~#4hPy w7ڧ"Kq_ȫ"]|oeeʞF({$lBu)ۥO#:=Wcԯ;v}L>oݹ._@,ߝV4R#);jj@;IbPbwB$T@J:O#y$;N4BG@JzO#{$;N4BG@hJhO#4x$BtC#f o'I?57S"ᱪՋP;fZQݙ1 g[Kn-еh2Cr3ZoD~:.D^cQ6MetOќ!"+S6.ʣs"Dє3 ;=J^0( qs0R]p^1Um-Чfy66v`Ԑ* fEq ٷPOs~$Ϗeicp]Pa(ehjr|j#,OehH|+k\f0{h>D;1Ԗtf$s2ҫ(;J}߿֞UEJ@i8A /P-ڱmOQ6$[a/eTU ܒ"[nR&DTW޺-πj +8K3Q,iQ|xn}G۞u4Eb?T|(5̷fp7NtvN"w FܲzT4t?ĕ *cY2| L9_)k nMU4D$់H1CŘ>u7Ep=ˁgdӳ4GDAkHzS^  9V@<A5]|޳`k# Drئ2SY_+'g@̤J gUCKL@z&3v+=Hi:&fŢ<9͒m6@ hr#sqo Y=~)yV,,Ϛs3 \H8F x2SeѵcJd&@{ +9]m̢BS(>Y`sH/3g'UM} y,<#%rs#O3ޘM{JgK6j_}FŅӸQɽf"P:<&(_{PZ@q0Ƭrf!?4xdAYx)O$ F9ЍPkN֯]+.-Ig K4>/*$ĬQ-pS \RɁr8 S f (#Br V}ҋwy *䵖:Oy' b}WQ^>/y%8x_"ˈ@ R|HOHS=6};Oev[}(rp'qirJ , ՜$R֓θ[՛}&s|jd`$*?b7u5WM!yHwF7\c3TvzacQߓhfIKo.Yib)hKhp%}g6x@F]OJ@cn6b#"I(} OkB5-JyZ!Ϗ}s 8P<Ɩʸ3x:-ܿC* lx[5wi"C,RmѰXSUZ_rAEuo.ۤfGVt-ڗ` @pg3ʠas3He5wWKOKN'Xuچb} 5ykn2aV)\ 1G^<l +94^&M <]OJ(}lt*HH eGpr'׳Je K3UšÌ5R{L26 f%eOqw۽Ҽ@/;{!aG.K.++7Tr\jc,'tdIϴY<(/A(^'+cQXLMp3iK iQ~ĤE7[j#WuӍ.EE9FQ@6ĥK+MD)D Zml;JdM߄YN8<;7z.@%̷J_T &9ods7N߫ *R$$al 2L"ط~ K>|WUx@'* Cm% `BsWEט6Q͆, lk2k0?5ԽՍtu*w. &L i\JwdOwX=)%o k'wgk9|rȼCWٞ~h'íw38-C诽9J?/R1 T9oV4joDFY&1!Y4s΀P]r[GW10Q+  $%~d].AHFg/rI(+̴@M=_>v-0 ;csbcz OzzAZg˜(U!Nf,;!o{pr(zƦ{3cOA/4>׿Oc J~c>d>x[yq={:lPv?/ڼM?Wrx[qab6د   if6tuczzI*sK+%x=#dnqkxl:`GuU%x/vn0[ybio"О}y *2˽b:|*bRX0h`$wUۼ< a,<=3x4÷om-E$3;߿G1+}Q ^ZfJ-Dx F{eKl}v:dE}>0dF΢šr F {Ҍ{  {dCLIK,12g^Ln1XެX|vߟr[~SHw&|ҏ}>WI3RИHĪ&#Ѓ(1'L)Б-E1ѸҴÈ޷kO~J(VW!1!f@!gκf:A~Ig<0U-΀O`,XH})jU8*Q|rK#!I ap/#M[$arphNgNOC%c `$\-z\ SDJ հ }Z#$JcM^P3sarձ 5C Z5 <.F@Cn 98q( Y.!2Ȼ.] vUB ,)я . 0EN$8!DoZX贈y se<>c##9Iš2qqFXUg4KxAh b3<&я-ˤ&Mܱ98T9 |BiL{#mNYm0zaey.`_Hbxu\Th4Jf &>}'H& P BueU FR4:v-H O (x)Q9doOlID4c?&FT2XxInX8߯ -N\m"E|[Vk%3,XlWE(hn14F\nQCH>{ߒO$9z*Ȫ/{d˫y1`DMXzlI8*"io. >݄b+bau饘J2\TaQ9ۧ'B/*a)|IScӸV r֢ձ|-xϛ:GHb 27+9$oJ; fi"&Ua^4|*s4Tlt^pk N-w?ˑ"Ɍ^'txxVG3x }:z=z^obUoztgE/飾h":{dicTt&1Q1YbKaD-y=(cX,%|7^¼,)mt-RW%;]vw~2YUyU݁2V/fV̋n˽!qഘrX?6!NEMf(4R΍UlbEJa$ܥ{y7i;5DžqR.lr!7\+Njp2-t6<ϼut%$cEUXK,k=tChҏ&9LΑ¢1O81x1|U[)\DC0}7DJA,e1bsMw=W,Cτ!%z;&s2`JefN^ {+X҂[J%Br3 $&vޏ-.p/̴!S 8\H3|>S8zҎ{a.YK$&8eEW[lsje3?3ZH! ۧC|ň ҳ*,/w$8* GTu3OFP2Ykk% 뚆PF#MY](NVRpmyCQH8׳pj9&\S -jkȔL}wuv?[[?ޮ6TiY.͜af3TU ~np5*)w GR2Q`7O0>bU؊>U^# ̀$ )nM`b*JDØH|4_:c'0~sKG5ZBXJ]ւKLxaL&IԏC=H/+sl 30iMDp]b`P9EM;9RWf/l#6GN[u>Y_{S4rmDQl!8ه=fMB)395bYUJV fIea)lKG$8٥f|ۗ$P<}chX4ڰ<, Đy{# 9 쌐|<r!17C(Q~f`s*!- 7DHiڵR[|\{$2`vp6&oIDң|u2oV,6AnwHto2TTY!kU6`1b_lXpF7${9 j|R bĽ .R\p2%`_[!buúgqӫ;A[i*kؠEFb:13Vy =-1n]wݦzETO;DRn?FϊbhSgq y:#T֠LÕU1H>MtI~zП*@{#!O0 |Y3f3L[xVœ8մj\hطuԂ8Z+D(um' Fx0S)1+k5蒎 a0MW掕AoDwo[1eOC&= 5X b\J޿G߼0ȂTYャKfn2*Ш\Mpƫ9A+QX.p# 6 e92B;w['}ziVroPYaGrv*Y gJLiiLb+YjB"ݗzO?xߞo vYZ'_"-k9բFOJ>(*>E҈w ֈ}jW [ e$-Fޢ_Hpꚫ5#p^88!W#[ n4!Mm&=6i,`%Jx" IG*'h8uNiC*df40b$K#G-5swB.kͯv1 ɱsvQS3 נ;nQj" ĒG̃',.z^Hs^&ͬՕYl"LZlPzb}~QYcӡ[6}!ѻkon^e)Cg^zڎ\GƇIw@RgOm_x{z8͐gŢ&1Iz s|YCJwEbX?ܼf=v[ffmgŸv?6\ۥ|X:Sf8txF5@6ET"8ȫEAko(?N^CM7HYC֠ï ꇛ7ilWZdP &ȡբTJ[>xa[]A,W7m.[sK暅\d ?~HKgb`5Ԑ@x!"W)C>ls!8+^Vא  v0 kdj~LE~F%^HpD>(S2$8.~$/t: 2k'80f/E[T ;cIxHnE2+p<!/\ ,ۋ%1p᯺?5ihWvQD^+H|yIc[}cw<ש۬K[bx$1ӯdy#|#[G /~da\,tw.&aY!H#bF!2PΘ!(pT.jݕ`(;kVhfy2Bs6zk^ *!SIޘDJ:bsK:_aCpqrQce3Idf~Ѓ˭.NC-NvNv`q 7u^$_3NF/JJcK!{J ^lA.b$D|$8zf6oH fi]`^Zx 1SL:b!+1XDUŒ[֠LDs:wsA"YU:<ņ05^Hp`V`ZrH0^޵6r#!`s!\npفu`E|-i&WnY-3ܒm;o3y? !댦W} n|M<7W{RղEXcB>uށ"Y7|Ws[pǙϳHMU_\sq ~ue O@ R\ ?ǍE JSI%҂+4H BM\.~ْ:X]391u ׯ~iy1N7 ъר|?|<#k"kv%h$Qϒ&uK{kO2_(w8 .S8 Bx|EoԬf?Xsl2r?yeW~OH~\~$7~*9͏1-{3^ I|%žB"n [4J4fql. hSrQ88M  `].#>l롙Lɏ {r: c HztT@0FqGUq*p='7Hdj0"GaI1A oz6&c1+'@\:4q%o|WVu'3=hNph(FnydmTU%=,c|V!_cv+:kl!a1 eu;-O9+ ͚+k?80ʮYL\^#b95Nf JSDܺ\u>KyIRUWHYm%V] TMdt5Uþ{VH]@+RصZn+#н{_Npr+6iVbV{q'_=b~W#(#\D [8?r.v.z6G%VfW5R͐U8U lM*k}5…HTq7 qvu=ψ{-x?*>*h<{?<>ͻ,)UǂI8+Q[U ϓS ړzɆ ٗuP^܍`1s1I*`CN:r~v Ɏ{_Wzx*5<Aզ>a5KO=a=$3aٮ@y'FP5o?ߨ ֮sM􍪨%>_~֝A썖>)VR E]=o$?O.V3yQ\ݙK_@ZMQ+Pt)%t cծU |8=HDj WBw[5{yZK/7~"&?,;A\Y4U_.ڍ1jpF )7ZzgR ) Rh%By"σ$d#|idӊ% @3\' ?Eݬtƕ7 "je4-XByJrX=F%fN.(*@h0ԙ-iE,i@P81IjdO]Jէ[@\ꅁ#k8e`\ lC0@Q&hR*XŸaIvs#7 +j VpвMn] X &,A,D2pnh4 @X2u>H{75?x0i<``^`^)Tڑ :4Kԝ:3GDimԎ02Snծ좼>^o.Nk-O>]PW4FIg0hDސ'qYi@`o>'| S`A |>͠=1/2T֏^|\uV.{$$!>"}`D_/_%ȝ7n^cG5`?4:JՆVGڍLf2?$ߞh==Y[|9.-dl '1vsTJ  kRI G^ $coAMO؝8%4_Pm5k4;X{@|=G4_NǍEFDGQ|_./l: am=-Y[ Lc_䑯hJZЎ{kF,nwHS,%)4#$Ŗ*ͭWH;jzj޽$6In3$Y41NZF@F'9!`uBh!Kg; r$|ՙYٺЦ1LG )<. `4 T3* g"8DCThX3hfSԊv(,Ѹ6][Xq)VaשT_yK$.M@X&i)KeFxnGB1"2gkhF81ZKE͠ n@S/&Fo'0EV .VQZ-kwRݱ\?gI. g%va-q1 HΜ1AJ𽄣Ka /|9n{-›Ơ~nDPޝjFM%:ZκZ.I ኵ֙,-Fams 0jEcUl]|rL|hیa6&%s%[C/E:\{{do7*pc0S-m^sU$\`U"P/;d%1,cf8D= ecdG10ȥ{c%_Eз˺ b\YϳQL}=4VYbZX^cc9Y6(oMrZG4Qv={aeEL?%Zkޙq8+PgTq&io\1kQKon@=$ AsA5qɃD׷'noA GGZPBŽ zv#cq01T鴳a ;Re*YL%%qZBzl W6+k![5)ّ<ƈ}n\ܻ`15?5=ۉyj~ i B<9" ; AG)f# ? (`3^!zS*QD&@ּ@:hec  q P~~l$&3@J0m4 :PiB=9X,r"hO IQ}ϮZ Rh#Æ9DaDlJ<#K l¹IHd,2X 2'^xP +^ZI% 3ɻuw1I4j=&OŚ1 nVRAryCSmԽ$Ooc-xD_ju]1#1 LL5z7Sk(cOnc $F Nrxf88KXJXʑtY=8[aw`S=jVrԝq.EcK/ fhl0A2XiC* 1"TWQhs|>ϱPJbgy0{mRD$ Ŗ8+<Ǎ疸?{6>MeT&3&S)m]dI+Jckl&Eͦ@Chtsq4όAz@k]Lg1H /ge*pӠJҜIm,w~3ؤb}LyeЖz(\3 mL_ǝ:@ff4goa1:5 = CxVcRqAS"lLP9طQ*H] M.4;y =!?ޔ`0N5&WyͨՂqZ [,+6*< &?^ƍ Z%3nT}oYG[#Gx%:f6¾250&1I12'夷z%} Lފog/lwL!&0֥=Ybwda;[bv3xBnFuAN-% 70#g12Ltig؁i\ôZ?z)NPܹi+0`3uf}jd;|v;(F !x"ޣ4LjYιWFeQ(1`I*dwR7%s:#\g,rO'>8΢ZUwQ'T r1J{UkYsܲ&>:,J0)ɔtvo3vxiFfP8Ƶ70)nPw4]mWA8v,t#X3,auRRj_YU(@"C t<_*m>anխ٠5_t- 1Lcq6glȨPfRbALQkgc!UY˪N.GSQtxQN|A9ɦ6+73jQ~*tf lx{y"o+^4e,%o״i=m=2tP>LU5Tq+dQD'4#؞?܄\A}@s@ɚI[ ѓ Trڡu;יuSLjէ +/t]JF(o5ǭqN3gCXㄕmLv} b;,DV4X/Vx`J_J[˦(e֊|lYCЫ?7p g} *շZOolٳ} ytJJ-+IS`[UNuäAٽbPFWZy}YF>/ @e1&jJ|>@cwׂ^3xrb(/O{=lPx:F!09-_ tjih0]ǯ{lE7]CVGbE͜Vː<[_cioZw;{Z GNVß`5*5ȁ"`p,HP8Z@ޔt)l9 7EX! {cX/rve|UY{!@iR{XD&:{v5c$h42(7!SPN3Ҏ>doH>mr;KU^S)]?c!f_S%5(s|iY3aSUP>8݌R}d2 :ɯ$4#Qzb،\vbL[z\-/>Ō8G!A9$tG,1[2و%鳊5"XP Jg yX%kYDa;"L/PO߆ PP-1NeLZ``ׅ {:_9O|'Ճp_ P ~# @9-=P, o[ aM_kʜp= x0a/b/-Urˋ6\fsU?0/w hX3!j냺bI'3r+9{ۜ<ŗN; LpvH[el0aIuU=TBz/]$c<<9t>hW視hP)I.lxՔUN iQ9,7aܧeLpJKw({GڳW[j*d?dݛg>si[)tRZ+-=2wѾq0UqU~eVDqJو-˪"?ʓgc-[ ỗ$!1YK8T?Z[e_$lV fX˚Jrh%e<S,KrSƟeLj{-,8zS`+P#0B7EhcTDVC!/ziX51j%Ez 4i ܜ . s<Ǭ&I/0].kxB8_ȝO"$[O19φ{7i;?bNyW{ +e&?Mj}xZ |HΪ}Cy%^+x[''qi:b:Y]'VeR^*G.G}n5.&S>l.yqEQJK-,qްXqA>׸Gާz_ٽoÓy =x\to㉺|&7ð/ysE%r1HaN"lySxT&'oxILd:wSs یGVr{w<͙H#'R@Qxk9T[pSކcU7= ՖMEq9%BiҐNK๔gSZOӞiu-&b0oE(JLDH29:ϼ䙉 smle!MK/yL zIinOՆ~.xltg,"_q5^g햮\o92O0U Em0`Q1Hf:Mt2 ?K4L)~;PR5qՔ2?&rghu/k?A`Kp"0w屢jOUxz_"N@^Z*p}}o{wNFZFt^5Jgk|Nu $&z j%RG߫nIɛ_$ T bd?MIĤE}ՅJH&"1@UL+S&4:W׼-;JO Z 1R$:Ad 2FɕWw5^qmUcͩI[L.g ̉;[ 鼐,0 <'V*w{=@^ +RoܬR0yH<50~4Xq,7:}T ׼-x,HH#qeJBD%V-9OlyFUռ%c+͹έ5E%1=q;-ⵚEm!_|/kI=8gA丸922@H?hfNd76ym!?as,a-sA~t S$;Pw!_cy>*krkϯL Ƴ S R8͉eAʄ\ʔAkT5o!?_/1a"G8 ).7}T!ռ- +*$%c&!yyk2Ru^[y[tn*:4#B K$nDqIh W4۝o@rzۂg\KB7Uvzup*j`1}myI>>.H7`@0:wKP5o *3Q5%ʣ2.$Z0d>*j2HnNzɏUq YWx_^ν wy 1P&J.&Jmі 帾fWt`7@炜K77P,{q 3G hb)I<1jA9-g$ͥ5$^7ylmuǓerƛu1O$dS$ Yf?򮥷Io ~sX3U%$U˴dUJʴ"f2 3/BQ3閘aGbSƍ*ϕ̾ |#!QdY::2! gwױ T ٸ qA9r;$IN&C{ɣ Y>V-}Hf`Y-~P;Ko۸*ݾED [O7],Nٟpt/O}?(e~Ǽl܆qZcpM:X] HI&kZ\#dXz|J<*`~52Ox45医Ю9tl^A% =<*Il ;wOx2kG˦eM'Ia:zxDZ>Vb:|{ a.YsoE058YL=2# Y>VRN"x4jjk3,dUsV2QEIqPɳ{ &J$:`ݓi<ڒ.lX,jQ\Z*xv)Qe*fAjGAG r&stw#˰vGL) ]j Xrt! [>V`96=6zYG,Q z?&h4"^(|^3筂$F:,HbͲsOIXۨ%h'tANJ\';7=qod l +cRsA֊, ]4CZ)}5OmӋkPG͡nۏnTỤuR;$哦2`d)/mIz{@ d.x7Oy; TN–?Q| w,,dq!;' /HZ̐-/oa5MF(g`jCDt;" Z>RT8:FvgJ!gF)89)5j=2ϵccE[0}X~4(bYsNRVJJƌ1RdoyRvspJ^KzQ-)dXk=nŋsLQH1hH)p C@kܺ"L88iB.21)%wC @-., 72ك _¯)C<5Ԧ`%n>η~ʝkn\6f03U6𨸖Q`RN-HhűQ[+ϟXGdJ1SCIhlyu< L6] vvnm-MnS6PYuƾ nqf)BCo3cE]Мy]ͭp/\ޠ{=aZ> T m^|-f=)g,0hG*xƑY>V"y>}ss:| UyI\k9ۙgk64)h|M; 5$Z@fWRHQ'F|t~@Z"T:PX"NˊO^,pohM"<`lZyCV Z%&kp'{S>cE.*"v$*DD\,n贲kgz2>}G zf]y7`$3W FeE{sȽ`家 ZtkUk ^ѥfJoo7,UL^44GIǪ.2s<QBA. pzPNxjUK(&*i<-@;|5!-E,C UЍ yANJLz(𨸪OPזJs*,Jt0W& !yI#@J-OBOrjczx\#Lv6mg3![t+,dy xx:h\Ȑ/X_i b=qv/ Eub9+_qƽ62M mAGJ[:MR> 8Z-!]faP1w%xM5f۶r@<ȶ5i2]#/ϫ5)OQѶx3- d% <=CFZ| N{՛nn5`ms w7- emm|-@|+zGϨ-+fEJZƯg-OUbP ^anFu`)&_Ƞlwކj5&|)[ӭɅj Cm-Z>VOd*\jMKrl@Xo+F{%Q>..VΧz  uB"|n 5^9-xm{&.VJã- Ty _q̛YzNΠ>Ѕ&\5( Mov@6GŅ 3y:@+.tX_oZ<ȸޒDԃJz%)8|]^/WwS3?fU{.'kC>9 KZrQ[!血x(L%+85n*{!U,Wφ(26Z1xNً~m4[o?-7ە`>͚h\/YōQ1]oVs;9Ml] /-1"Q\adBM>dhd`JGTkU@]*bMHV|aL8Pk_ɄMgN%UIcB O}b.8n] "5CFt>S(Eiʅ =_m Z>VB/Of&ڐ?0:Y4,C"7J 讟ꤵ&z+QK賽J Nx Z>VA =βyDLZû 48+3 v|tj y_ب׶AVl!;!]uOX6|9)j,v|l=}Xa /03hgNqc;yjH45cEN+]^"}FdMJȹrŜ9Ҟ!Nt9E;ě z9G]\*juĩ)#{<JVT&V|6B6bUsZEG#M||Ik'MPpMd<-\֠TQJLV}Z= )wzxԜRB(DgoBNth_!~}~S&_n_PVGi1]4Avm1 hN(5+ IХ2mP;e5vhS:bCf9 u%F P?CD@-+uVrm-W(io8{6ay6xEll&, Q?him- Z>V̢*|PMꨱƨ+>?o:lYQlm6 5-R{(;Oy*%4ڕcEͷΑϗf;`t-1Eŀ! cQc?\twWN.B9!\PXLWkl>o:07M"6۶1g&0C@=C|U= k >Yi>ct5׈LUcT=z'QaIZj{C͠LZwY-H:}.?z4|tI͇)"(9 2gU,*y\|tUUaNPEt+lPY@{Y*hz~,v6lǞUuwA/: _(Zо8k_Pq$aM&!✰Syƞ! ftQ!5JKєd]w(\`$~ yr| 6._tuG:MQ3i.`OxQ ޞr`>9? 9LxQj_]۫*8 uPw)&iUb. SP)^g;!!T@[ph& Qiv*<ԾA L"ofPrG nN|~ڊPxeDiP>ZoAfj٥VҠ;<v,| _ kWK;WxD`Պq>,R&= ϨW\Nj>=ߪ1Gw\ʥO\/.ɚ3k+'+iaMQk.%_)ѝ"!-g8qsDːcfS/17fR=Z,CD=zZ|qG:2W6TՋ,mQ'_a:nS.zS܎M ڞN?.]©κs~Y_j~?[bxo㚱rh3h+ t%,F"_W`=pGNNzAmo4[m 1|wP~w޼_Y }=9`h<EAN8yN |+7we5 xS+" tp{wc~sWtÁ"柖FϦ6U[d𵈊 }zЎ{Hmǂp}:J Y=ϗ|=`LJ2 #ƚ5+tò(mܢжCaN]d˯%w).]|cn>F8SZZ艾ux~HS+<5Նmhx7dkx%iq@&۞릒+9G,>fI.RfXj.R0wMZíqwwBg Ņ3boK 8r9(d~+SZF ;C*@NbTIbB+j|)F [߃w&PP6\uNO7Mujעb{6KQ"a?""j=M}4h~.@@' Cxa4Xb2ˬQ)q"3o1k WȐidAhczT'@@4>qd5nmǘ= kxOxsPPTuRb I:$6Ҵi5M_Ǫ:Aߑ>C"O# ۝mlimdY8xR?|R!>Р^ס(6ɧ t avolAV?0 QϖX!ᓹˌ!njRj AH־z]S @p!6Q^sj$h WZӗ~./PX IˋHD8" cG~0`ECߺ(:Lϵ /<ѴOZ44y-3_9mh N0%8jx+|ZpvE?ݶ%nQ#BltY(AG񂮡RT*)"^HN̻`<Ż9bA~7|xh#h瑾ƄІCx4RĴ`Hv~lo2ѫwG D^T&iMQ.1[VӶ0r㽗`<%NS(Pu"x#1:IiJNr4(4QaRFA d!r_>R*`f(SMmlt 5:2JYuc<+ =p2Ʋ28C(ey|6!Zk_'żq6 z`A U)M H.r8"S9109s/sc"mHJExͽK}`! idx2oL\@3Pt EYTӷiK(q4{@;.L)8aDIj@k*"kIޒwܾ%;~@1s .- Sػ /?qtC,b<0ʓ' ZiD(agv>-=<%,"G}TNiH^IsSHlѽ &ɀ-%8p76kp{d']RՎcN$ax fAd` ƳVt-qʼnWI18? Ͻc؄߷$kҺ63 1at{ #0kV _-e+$b\*#H(EH!sxoDFsaU5ԫV6Y/2@qcn+` ڤ`6t%X56־bdF'^'5ݾT-Wiz>UUh~"umY4=OF*XhׁR{˃ J?6J*M08!۝ lw> Iv-}y2JpTR1So^Ӈy MءѽI(Q$b>H닯b_p)bVE'1JGʊJz%e R +GR~zt$^>rَlπbPIؚYs6)AiM&ksF̹{Ǹ[/|:{o״UxCS.R`F{sh-mA`HVy, V'BVJ!IQXEkG5%7Ihtu,M& RYw +QK|҈MaIev%N׼0F:7nGء{9os0x8yKFNXYސԩrEKY],H 9@O&j J1Gu2'U # p1#=e\cIz xnu'*Y'}EEO'%ı րF0΋6d౓7zͬ1ﳻ٤gl~V7jhln*n62O8v)`%>׫M~f-gk{n.s;&|o,K&{tu(_&bAs=X:,DkcjXsGKZKb&F N;Xݷ-/Y^Dm3DŽk5w1}'v>7V]-\"Zd #2 y.I+Ox:25-f%mrA7xҷWoq5IT(s@ǟgO??:ɹ?_GO1=oigҧm_rOS/xGՋ(5OzflDiIi?jvw??:/Oz32k_-Mor|9=bVMz-M#/-}g9VQ9ztᵍ)ףrǰ /*TJC ^4*+[ kxb M M=O,q@ϯ'ia>=>_ƵՏ&r"_4kjmHC'I] %=:r7_r -񋤨 zg/wMrwy҃Yn͎$6}9ջX">h7hxMέ1i-h$5Ǥ#Tj&>7E**`AEsZ2 o?QG<7 ofVLӸ쪵j *.(͢x/ {RoNaA7%eݒRYAfY|+1KxA%Gq'z썍"@,'zA>=^MTM+޶\; 7?~ Gd\}ܑ s|ـx71jbx#X%Il"d!bM2@7)nXwgmmGJd1xE]6W/jݯHו*d jv7/wk}]_kDž}eD&bZQG 4\PILdR}LXJKéVgRKj "LDjWbNo}||sdRqګ;QBoj5Lb l,>5vGq41IZ1m+G_`Ye$`APYGF {292'6mV[I_'򜥢N(>h u;gv)تsW$O||@;t7D!`(q@0&H̟cgm} ;GDEAom6Qc M (U)xIv 2gDu!n@U,JzBFOM)$UV77 D< #1(}dq!Vޞȟ8ٮq fkm@7jPصUs4f\Z1~ǎP.l7"5-7/OAUyl`(l%kCKw`ߴN ;oGjhQ;y^؂.jLgҁ9qyxʆ l8f#HHjE@<.*>݊"Y]z#8U] -QEuvY8J\Ppr+<}E{P=(qvøΊ/rF3l~ 5)Y(i 7o Όwq(FwS#Fj?@rFj#c*D \rmPtPmVQ+TCBܶA 477^"tR{)S^\|ZU0q1&,N6f,0Z̺# }C6ՀF_iN ƤHn-\ω7srHナ˃~L5Q[r.Ȩa ɏf[wm>ة R?n<=2\"lR vqZ݄wv%auw?K?t,"+1kUe4+Ll rEB| Iaߗ; 6q͚/݌ON.o~<ٹm+/6˰r0J nm>=/M"40ր|xz?ssymsyge1ü`0fUP¿\eX{wވ(tO7?qawMg|o&:R)3$.l_4n=a\ô+B@]fiw`xﶋ=ƥ;7k?WYҮwz6G@ZLڋQi@{AZ6vy:I;GLF|z/aE%߭5Exc>.~OO ͫ]GA]a9(A"H^P͓$(@0gCDt~hq94/-l3%( ) (YDI78v]^B:.԰Q/Ur&-8iձ )H,PHp Q(^)\系}K8`8Up$X `i\ʫDR5 g9qTN`fF28I4FBuU9z0{qba4p}$JXPU6(Ep-O)< ҡFxa5,<B vϨ)6 RJE*d\,`-զSǹF4 C =#IT# a5`:ej@8C Y) 3qZwXx{ݤ&T(LAHdȧFC08Bdr3] `),Fap2|6Ʊ$a1 9ORE].eW!0^ c/WvyX;cmyP03dHBe-ǹ_A$yk&a*~%t[ B֙uEÈK22ư v*Ĉ$IaeP|t$x .iZ!b(au $Up<SؠQsmڟʪ}NZ*pלa *HbLQ ,D)D}wy 1T%+ZYDζZΫl{g^Y?v_e:kJzh'ëcVcdeeIZҠ1:kfd( ŗd'Mqn{[BŒd0O'Lmt?Ϳ Ц k)sF([' B/7;mlsv] v9tUj:@0Nq394$:,s^+53x*4}e4NFc^A('?*=DOH+C08"?8k5@RrkBaUJP,HA4 #۬>-q^akAR{VwuxwxFap2ü $i> O>-KgUBBXY;jd wA4hgɛߒSKf8ﲒw$Lzh&3͹Q|:u/8\0"sz[4*wYN Fap=D`5Oj{{Ā|VUNX'&UnXL°db;,1;QippwIxoJu~$5( ypK,`8Zȧ!})/50#1('5]ep #^W) W cXCщ BE}hua@p:2I>"`~YP`8p$ĵ$pHUDc滦Qt~w?LRFap Sϟ9B4@ArlLZI 5vǦ5Hs {\>3|M96#r 2€ˍ鲵MeAǦABK%I}HY&2Amz 9Eoi2XNVnG]F+o$D,ko><{hH` sL Ldȅtt%q߮X3ۿre%-,~[F]kV𪡢&QW1 >H k08^*a[cr ޠ{hbl1A-)56T.5'j!5ET_oҤKA-)9- ؊y/GOex@ćzDp!}@$5 D#'9|1X9h&#Ze"g_R3S -B3yzv{F qni1k~D0, H ^ +l95qƅw*1 \ ptQ@kɈzh'vdS 2q>J>H۷J #Q fOSb5"rfRM"PMv@HʶCQx WՒI0qfv4!9O_ȔsxIHM#窢xh#+^ i7n[5P1}abtmRq9"ϐR)٬|*$Hz Q ΅[ZK~L,N )zp=56 #3ŜةJK9)\$Y<~M0F4K AI/NJr/;aP24):Kƪ70Ki$nJʇC08"[==%Zdp*':Gb*n}; ,չFap4\8iEn/ JRh~*9r\ $FE86ĊVߥ{10:/ngjΊd\_)}HsuCRft&-6og Պ.I7ĕJLRRh1Ig҈H 49?ĤC08r <_k )0}8]( ŬANrNrMzRa(FC08O\3*˜kK#U*jtz4򮜠9( B'y@ѧFap4Χ[u@$;\-QΡ r']=4Jc?) 9U($99U%=iWdpww & J?dH(|n67¤xld!c8 pJZΠ:TfIAKc͚FaprxN'Qsml̞ɲ-PDah2xݞ]Wp^Ac #}D , tJ$?Qejiٻ߶%z(0h+ 60K[M)"e}~>(YipF|U@[T,<*s8âe0A$D!.cpzJ?K*r`KQU4+oG@4,̞D i7Vo$A C{cM_ib1 gPGp̸/|MT-;,[Z%n˛fG'*iRY#@.RbBP.u # rB0a78y7xNc!5ڱ&iYwZ߳@\>@mvطpsjF@3z@Uyg(f "ӑ;JqN=}yVYٳ3N d<E=`Eǚ[߳8z@R[sNuf`9X\􍞺GVuwسpz`P(r<~YVIe W>x*B Y4oۏvF'?B 򛉷7NO9jCcqRM9~M2p.Fy}w v3 nTomd;~__^i85#|TJW'Q"V{uYcJV%G +8f f**EPAd#Q@udȣg/W!EJj3F4O"*e9ha%ӓ?cnſ~|utԾ=z{;BzF o^rAvJ4!M6u7W8ݞJ$c)XEB:aHpP V{I99TgH\%Tjs~QkwPՃT;rEc`VUV ngl9%yc3)\ohPb CDoX@+9w=Cbm(;֚?h7U Ҕ׃Krx9@.*lL&ASK:S?ꛩNцO+ }6|-Z—3Lc¨qCTOҖ_qooJ3-~94h_zjp{>tU=ڡ۩ɋժ y;Wp}S|;~yR1 p,&p1"Pʻ%mx[Dqڷry4h7Uc$|貍3sLjOz_$f(w.OqiX47ǤOÎ X;))`+VqQacɲ?3?Z^Y`ޯ5Qnz軺 ^|e[_K%x9P(Q6Pts炫1̽;o+KuW_9`럖km\njP&ZNӠ0kE=HWDźo ]8'νMū ]zك8o eͶmLOd1utEʭ}z8ٜf|_,d}v|ȳ ٴ]_ǫQ9 ܇ҟȵ6 PmjT6@}P ۷LH{QT/8^;{ׅQ*W de_ڀ>kJq#oX<x<,x<,LJ!%z8uBMRh@ ,Tس@$md7RXb!`G  ù/{qx穇cBH OtNH"W5w}WI~%y\+o)p)BQǘD;<4Ŝ)j$M;q_Sƻ}wÌ{>֟a6xf.Lt ?`HL3Կ#qT%a T }jE& `F甫Cwhw^ !ɪ ҟL!Gx]b_ÏPEUSM'%"dg1Xy#?CeP*eN":{͛j=}:^W f傻:j183܁jhVh`E#יu\bAP3oyCͼfP3o7P3o7f!烔Ԃ%熐l(97 %熒sCɹ l.Mh ag4{-a ys lsi^h/" )B*p"EA0l8e:MI륉:-Mt%t<&3a0?:Ml;A$+&13:8̒Y˜@׵zMYׯ"j7 l?d CMg;0PU|ɋay*{ʞ76a2eaHT&a#ǩFNt,L"R20+N0+'d&9}l25_X%m\\8W{݁uҥ.=(rLA_6Biyw@{?vA/ e0Oquays__,ڻbI(N Z}1]t6dUP2鿮Y5~6(q,iىeYPt<.`U6νVo_WZ=o*>xJtbZky`תW="t_Wd?lܛr縱}6 مiԞE,On+*:y_r|ǫtS&ɮI'Whfc!Յg 331gEC3󫢋sTBq_B?}d+~!I(ۯ8Bma,6Y6FEW!9{J%dSfj!fkB~0:(7}ߞlcbeH\7G@\,N, SFzYNKL;$k'ىg=M4?Y]J_&7M֥=ع")U<\׵{6Ha)e1R,(V'ThjCI;dg/er}>jgCs=/dM:d<gyN`ڃd3vKzi^I~:ǧ''Z$Y NJ0'D;xyʐ2։l5H$rEx9$DiJ(^8OD#'5NFLyfv6z>Nڨ|KK#Uc#IaY>2]%m8 lHXfnO_Y.Dv4n=t';^8e-9T(%1Ð2{mNx f8U> !-%,q9$@p!٬uc6 bfk:Y"'0ijTDVsfQӠw,aDx LGgccP>jZNY6 ֵFAi"pIEPoDvz>M8N#$+7)!?P@z0>CsDr r!$t&ƾs+AXi/?4ݐ1z?'o';'v"ԖٝI sMtirG4TޜB*$]d6\=E{mC$̎p^9 )(WIv{\9W>Z`cb.-ogWz b݂F|~ߍ64=+|4N`6N1ފ?4*}Fy};%>{?7瘨=;K{ xvbA,څ~nOHiX \.{w(Mtzd.@~}|W†۳$>[DakB\> /En`UG"&*JW2!Svd7>vUUfm*;6mryo2'Fc`)`ʤ>J &slOP _fQh:Y k &!Pd Xc,1&%Sk'|g6=WerSX, 3 ǝC"MS>Bh!: (eU^Ve _&շ7XQ}}rkYݣwH?umf}0Dc{cjs=&ݹEjIxN[ ɰcz[~G6RvÇOP>f^(UP(UfU|U>fO_~.=J?{"_6Ƣ~? nl8 gӚxp$[9~伇( ek8]ztj(ӸD̺gZS,qoi$, Bǽճ[]sX^ݗIN;ժѥ;=Tiu<>[Hp@* y}%/>kMc´%J{c&-*P Rf1 %#_,$ Ig}sK/oގQ.:F{-@2G0ɪU0'aͻXT2G4jfXrf*tvXJ}kڧJzKu~qj<1QM I (% nk5ZXI;9 JiyL3}v>*%l5$VV Rx׊c 0وÔ ) :$]::;*-)۫gAZL8DFSPI㼎V(B1+5^Kt[(PlnX}hݝ.|t$Қ{iW]&VF+UH4ҞP-bVKi ^#!0f XP5犞*gUӧe?bzzˬujxBg*uoimނ !5b/.nGzeQ"[cmI<4z C.u+ L 1%8h Q~?9/fz]v7xhkO ȦCz~֔b 7)ΏΫb;RUI+V1̶*w׶}Yˀ1N9R_۸1 `%%VJWj6d%kp@mG:}){V: :#v8o N&iYi02%Zp!jHVofK(DY*qj̮ EMG|=+Upz.q\) B@Z@k/JԠp1LshB+(mXwJl/3PvǸO~МeTr/RQG:`((+9vKQ+rQJHq&3,ݡ{-JA"UܩDRkeal[tO {-N}Գ4zLۋESIQkY0:%i.#!Sd,d TTZWxKBMoʎlWdD2In'B9loN;:X>kgK,py,4$֟^˛;?K/@ٱߘ0JڏQ% ǵ_$FTHw`{ NǩL#ےKa+yG AhwgEfGPH;x r7轑AqucŮ LFo:n[ccL0U%o򅟴`%F۲TtA6J`5zϸZrTn!"rz9l|>2>WDvv,wy5rcDoѥX}qQCn&KD߾ ?x_ փ(}ӝ(ӕ{(',S@J=M Loڴj2Ӯ,#DxO; 0eb( <$-t$}*y~7Ryq@:<@l&F֠IP<}'QN3I# Y2d/pwy+ s46yg]ѭ~xcJ{Ҫ+iǍF0i 2}YCw_OQFoɃah 땬DF-(x#K&vܖ3Q]\MGÛ{cy哌uzC{-I;;2P=b䲰WiM#dٕxܦ{%6x\,RGJ/Wm1KP!P8Pj,T UK]EZ}[|PPiszx޳ۛP :Q/Z624D$k@!Kܺ1/z^`hO?Ǽx{ ⰧXk+W bJ;qdˎ^]w%oݗI讑N;;ժdM645MCS㡩`CS㡩T M;uV[nCS%Mxhj<45~CJB<+EJJQ?A\ &HK#Bj2tJm a7xAEG>,[ko*|Er$ʦ4ç:yXvOBr D`K)G yMdb{0(Hos=Zr6!I$:a*,h$X*$ED4bҠܑu4 K (wZ}Pty.> ƺeMbq8ǵ;? 0d8rƃ?7S;h{W[SCV!{>8%P0Cn(v'a:XS( Թ!̓tUN4i';ƞXFO)_j_ٿ?T ӎSNKN)9?Fo絁hyEtgEemgzY Y~Sd(Dd&hƣ%ov+TMg_EPuT5s~d*J:2 dm4Z秳b}L\Ll:Y1:MF[^#TK۩^jFlcy:3+IYT6 }ll3COu0` ʣN+Mˏ3;}xǟޤo_}Ͽ՛^?ޟ"E)쎚?D^Ҝ4mV H3[/~7]e<-r C󑊬.Ȯ€yr8mʭy=\sָY:/%ؗ+q'.(fY)ى+ߧ|8D"oq]nq/'JR'8Z!8E`J> 8 #4sih_(rJbr 揧2X$Gay~ g3.ɘ;˱ӳxSک0 3L<"s吤SHL3ᴔv8iT&#շ55aC}ีYݽ66?N z2oB[{Y[<+BFGµyC{/R* ɐT#2\\종,!'Vq$jhO-4T #y_ƛ!:"g#Ж( T1@x1 %~1"]@{kf@+ZSk_֗yċw\eK-Hh"i^#GR8a$~TدnE寖yd`)S?XT2G4jf8k/h*V J}kڧ3%%:D:8JBѦC@$UqF7v*j3V_N~Oi W:FZ_˟g@4`ranɶr vU#1%d#851@Lt*tv >5UZRa`n-`xG/8)A8#PŠcMגx$JTuV,Rq~~9]/r(\*Vܭ@q,m^i8.i^!ĕH{BY- #2xYNt4`nC+ט+|6[tgU2/ D$Y45`5芥q\TnXROη=MĐ~bl7D^NqyԿEDΛukv̵E JY< YqYN:+"(-ٜwY~ݰgo*64v")!ޚyQFCկ >~%' $4`,A|g?fѺϳ-F%(B0,3w&?3(P!}JVG^D4d(qn^f;W?Fl1S¿6Bs+nev05|ͩW-!(?gv?\_n.HmֱSKNv[$۲-[rLfZ 8o g 牲 910RꪾDN[*%?d]ꃪ:L RL"ev?N?{ى=l!½h] \`L/ѽ_Z%S*ec\bNI $5 "pĺfAΣuw4UMg|)ƃ$޲Ő[rJ9l6`/0E僞Z27~[Ѐyh>!ZsS[1|nlJyS.6cӏ5a}afEE )VC*Z/q1O+P&YrdooEG,=!sN25vLKAhy!(&D٤w`_c3hci3H@Nx {݋`aH:3 X߃BwZLc9䴠x,Ԫӊv.р3ʃ?/*< I̊]}Ih֡_ *.ފ7\ʫQ)؆2<7#+Cb4^ k^8Qر| TP]ЧWu|G i&wK_)ޣmЗ z@ x-m 7J jEgQ?W7jDoԋp\>ӭNE6KZJժLXLtwljgZv`_O~d{z\gn1p򋶓z z6\'Vl; !=PmkNԷ!_G>9}X3&ފ`M!ƽEW ,dR. {&ܛ no0لlUmoWgOi[gJ-akiԒxk6ZNHb/Bql#{O};X/`|s39~_/GPg5Ø˱pH´Obt_ 7<~QQK{nDZט9:NUzh|jo2h5n>\C|P6*vgdNƦy 2#d@OW2aS֨2e0!^7r& kYj[ 3Zt޿nHߝiO[QA)_!պi"IGB} 8aC= ;Iϓo[30m x$m u2i"s㈑G1-:BrCegQ,'oF 0x"_+3$aTzRLo d#> +0H8 az+Ntš)XE|/9s!!xj)LSʋᯭo hɋH㔃^j0c}J[) UR5;Bvq,1rM6$:.kX&rƣlj|+q8oOSƮ]$g$$Ǧ1-$Jf1[f_|m=1Ŝ ƾm}Ow;9q m=gz@_fKGC6^o !TT%w#O?WE!\-`y!펿3 8*G{- S߸:"kmyX)]՟fҕso+ˉ.p?zgy?rS|75wfV}jMwsϮڛVm- Yr:C}p6\?F/5w[4yb]aN_>zB+KYor'0ޙ<|9]'m4{&|X|YES)G V%q#>oHe+^Bٹnj>bN~;7~z/g bjGH]iҮ{zi]s/ggi7. yUQvÍ0!VJRb ʀ 4 U+)o׏m{I L6MW5kyR~Vw9wn7O~>Jmasagw]Kr}_ *wt{r.(5^,g T7JRQ:@,niEb\Q%].Ƀ$jmj>s\t]AFUݵŃGSÞX dž 0,I ͬ2 oH])†j9>>h.}϶|?b8j[9;[\o]!gL^7jZ|~ļP9M֘֙RSdB Cdeˤ6\gBZ-Hn4 K)B/R(]=kDnmUG@yl&?L9p: `FN/`SF#Dz@$"! R :`' he_eq*A!n5XW6ņ[a $RfmB{A*Q R. K26 D@'A5`dC9A$U*aAHS"±䐤i;%mQBJEKXp,]vXUe_%uЙ,o:5YrI۫uS^ZRGrZcaͪAjCXײѨcTju|mY1)>_ X(%G;^(kÛ|mOa`rH[>ھ5QX:Z,L̍>Ż އKr]xoj8^dOz#Л=UrV1#GPv3ڛ xt oQp\)WUx~\/erc1}cciس}tzDOOd>o@ӏ9S›ƕr)d_uY;t}ttS|o{h@/૳ݐw燲AxOXQ=xWʖx;)DW;zgͭ; [FkUߢLiGGV98""HaͬQ)AЌ C< !I(4`)Rf ٯIY)Ҟ=\[z!-m.~zzUS1& Ep)q!7;U-wݍskw|OW3za" tsv{YL^wѫQ.gZ?5h.g*TD($lu2"Sy. =uYnq*ݩELT䟰p֕±[Ɏ캋n@ŧʄ%VG.4" egPtZԮU,\avKҟˤ: tm$L:#6u  @pf  çu8C1 SANW&6jAw_ lD2 ZF8O!y3QS³zy߳x@tfO ɤaZT1YhDCxmy@dX5<]!1giI{#=/0#u Dt݈DDxgФO=RsچSDPVm.TdL?Ӑhَ{=kaxV(?LW_HD<~nNA9 &g _Nv*wS޻)ݔ)p 1%&d 3EaL9s 4`B":\OM{%.wCCfnfoy$[.藷o^7KZߡ11__W luWmtwI] Hh-w,ʢ_WJF b6(Zep^; w$(C"'`el1Ċ3hbYp";cś,Z)o$4%TҪ6$ huxZ@d,2b `Vui= X<z_<,&YX";eGN̋ i[$ٶ qt4,c B{hyL1IѱLH*аc= xX<!f:׻b*IyƓ.Cyꃹ,NNQX!݌J5C4i{J 5'$]B/Kbѝ(-XfXn S$NI  I%ɽ&ʵd&W!kkQTߵ$"|TꕉsxHGi SF_E-KMgr6(ܺ28?Κ:`4;vZ :=F?VytE({n xyr ˹6ؕXW~bW|A!|P[:C8Fr[cA80 _>aHxDd1ʹ1LcI N)W!gUY>Dp%!QŌsm7b1q3\tD} Xqk`ŕ$Jd:|+Q2uz\nvbv}OgIKLTy2hU1f_a7ĞnoG7U%cPC+,"O)5RHpk uZDY U9q@肼ͷ-$\țKKAQyk`OyZ ym9!* bѶ1'@聼X3{U-_(\FbS t@IP+<_|ho~{xZ(2 MfUK=8dž#DaI`mQv+jVgMY0$}ɂPIq׮5Myɴ(xPVcxcȏC聼*r1Jr̯yX ,0Doɥ*jTpyyhC2PINyӾ<;j%oIP|`C~.;mz0UglMC12!0ӲbN>): Lwm dC>jQ!ԐGd~ {^3ɉT$9k#Oy^x ޢ4ڽcLBU.a{lGJobERj`VVdm ~ z ]{|8 ]S{4Ub^LLn Nzv2??LU?)Xt[XEq~vzg5RdEՄΠHԘx`v#ݿ2˧'w@dww'qb5E[;F=JwٴɦIe=i0NVȪd 8j ^^ ٌ+J=!^YW}:v7f +lRLFj(Ar#5D.2.c@BQv .Z`.m*j"|37w/Hr-mZlV9ig +.~uJgu@Yo}N.-f \Ia#^ J&#resdcog'䕵;݉V m-sw{zmv]+Ggֆ @Z|&>(`.; hY>H&ِsƇ╩( I{5e{=e4Jژ!0ΩP kw[? #1P#SyU7Fy;7b`XQ`Ǎ Z}F *}uVnm'pw`dj|-.|{4zbN^lx-ׯ\yy]>cZ"d"*>D&7ATd%(j#yD$x 5 zٰ}6'pf)d<{W%Hj$Q "68032+6k|;a])/A߱tPmco ;KF~f-:ޕ&0eQavPGA*b,Cb7Zz Y*t&"6!K2vt,;AzpA&'gʰ>']K1wq?:Yb1-Ҩ;? |?V->`Sߚl ѝ,e˽7aޖe}j~_Exqxv$wZm"vGZLB`TvVk~q_(oX"E5;ԇiu{?_v;ËV5ԙSf$|+mX1/7wqM<_[:h#탡ܥ"4 5bw{9IE Z StŖ[UUF(,Ef*k$켜r.Յܚ߿Đ2ke-`<+K R}/>uYf'feBUf.9K[k5ldϊXX3 uc(2]2r~q%vѻ5]NЯg+o['k'jnfY6ZFUy=*a`xF2)%ҭdWh8Zb漥t'rk.A}jW{D}IF8yBdRC;B]+5륕VfQ˘ 0ֱa;S t@C":`UG3C]v&`b e1Dhw 7PSC>,1MX- 2<[!Fu״t+7Q܈νH9a#}L[sf7<\{4W#<c6LE4Yt?^y6;8btG02‘HCqztp6~3AWJ(_@'D9;l쿟z] 6%Oc[zj'Mw+`]eTrHRt.CA *Ȧ* %->ё7h]0oHԚiDhLqM}|%҅RU[z>$*|5RF%M iʴcZu: La))*f%YxasW`:Ck|L ,?,~蕕ri dmlNmxkf6㜉ڷ{dk!P*KlIE$c4%h鲕jbĨ\YA**XmmHV'fb-mvt8~GMm.pV, nIa,jD0y&Mљ߁K.&.D~ M)`!6$$k{XX'GR %V-aA&bkx.ցAXCb^D`6+|Qt/]Qz qPڵ O%ir0٬3bh60Z3=9].ef#}M.g||hJpZ_, lŵ_J%UM8wAG x;3sz2s[[(+*Ŷ3Y"kcHIfjnENjC!Be/څb-LR_IUYK_@E&v?JE"J& LؒG<}BeW7d%v_N:,̍Ѭ պ$߅'*6u$olӝYM deY  ꩘h)ؠb0d*r(9fLҴDk$*`kMV|ɺAMF%4UMF6VRjэ;A{*J)hbC ms dICoET -`,3u{ޕB%HTo΃s %E ЫĘ"ee[.")9")aK2Uuu՝Ebʽջ,oWO]%pLi:$xbD Zsa1] v_/&f1+ QZ`Zh'qU_+l(MkōD; D{ 7:Wݑ\uB !xm6q-IEAVz,h*GUY"bNdJ*Ȭ(fjNFd=Reϛ5zzj[ >mR$v{ (w[ %QMnKl-yWBtɧDLxwoŖ DzMGm$h)D<b\ 9SpJ/Z5v䋘mPՔȘ*n9ovru閒w9{|3* &a|ޕU !9GrFuH2*$X1$R*&Vx(rbFym<*ruwUD7ڠ۠_N:0 )<0G1β\Ry NBP-mc"^6dAɁN˾!KB*@eJQ!3Hdi/Q\ГX#ʶr㱎M F$AK\p^CQRţ,ŭs6xiqZiΚ'$/4&Q;py>/;?>'sv}(#piH>_2n=z]ZLzw2֩#9Vţ;suy>eW_ [غZn]L[osXtf)KnͪClY`˹y۶IϏww>Ԡ祖!ot{N57Fh+9|Gdz2iٸEϛN385p7wm6i,#Yk~l}]{nP髳;{E!\}.J~˹x4~P FYexɇ}ExrzC\_q6K*lN>hlx#f *Yh`+鹜-uxzg˧0K㛎oNoĥTL$АSfgY3),C.qT9* Y@o%-XV3e#)%09ԁ̬-v{sge,jf=ïr-sa;myg4TFC?w`bW od!?Brt?[ <ݠ '0SAsEbI(DBfVDyЧWoҕ:J,uFg]m')|;X~6Q:RD)5S5’8S  )K&<%*C%*~p~3R^D'IuQ3En#f( }PLSk Aj B+%$j"'>$dC"B|acDNiYm\DTEtA!Z6t~2ą|3"2Ϥ TST&1,!WΏ15e&zDZ+JRVhí7"( @Rj)ZX-=uԆ_u2W_EeLȐy4SUDE NrDQr9@?=+B½m0z\`H$TщDD X \Ma B1ˊc ɱlyvQ9?gFwI?? Sr; /q/o.r_e>re]ػÀosYpDZMО |UZhda awZI),<\q`ٳw-ITqzB_#5Hsܷ^e I>b>!Vy.?LJޝHբzmn%4Q^FpyOF.MK/'nr#kc>V4kݛ6.FooՃٓ`1[?Ltoy%'Qo8~?;_ naHR?2|D q8`ŬG|1ɖIJkGedE֍ZV4yq#y`^F S%u*&Y8}$?8ß5_oWx}{8DꦠS>t=x2_?|hm ͇ -Tmqq6|1neBV ßqY=O>pzԚh3zza+8mdRΧ= լ>+K߆Jh)Ĕ 9R T ĩv!Zěp`2Qf$ %Z |o>jK ^Ȍ ^`A Ɯ\ p䮎5(mvbc-mkˇE62uӇVW#p|?LgGP+@G J\B:`^ek) Ӂi]IՇ XrreE.qVqLe r]Rld#maOB,G;10BB20Pg($^2#YobA% 90g H\*|-v#>kSNBDuCN/l>=ɾTb +=etHY:7QxA&cq_^&T*`'_CM.IFlS3%_^zCﭮ&jo%\P (֛7WjeEy{Dڅ^9s9eۻ;} ӐeG9ͫ,Yydc0E4^JJ8%\b%0lo?22o#=i`O4]KZ0bsͪ^fQ:sz×}yd#cc.yGm᫳f#S)#O<NJj>NRzJl݁I!OWN{MRn)~t,_ZH5櫿e0:=^dǣ{ JkX*T 3HUZrՂN9nr&=cJdRe&84)M;J e:9Z_Z_jm=:Pޔ5=8y{ ꭻJ-sCQVE'=azLVzvٛyap%&|xU쭪xxTV*1IXY:ط]ffӫU:~SHIOS/j/$bS'!,{*EѸOa zs;,:b-;|cY'*_Ѯ>e X`XH4ōǣEvRNG .DKVbIDvr%\Јsfʵ!,aP)+=yVrBV預wId:ܘM` PuٻTBYrђi J' h~}yl퐘ۦXc>?q*oyGu]:h*y&\w雧!6+%xƒ˴ Wvܐ{]6K2I%O@ dLE 1[cEXLR >fD)Jʾ%*3.@C&;;Iz<{$SN f~T+9w-KORy'9Q{ slS5,zo378I,lӢ -av .N l6@qt 3)h37SIR)_37z/sY̩K=eyR`bJu ]CU/W2:Wa]wyZzYiأCS!%Z\ 9$L"My;VCsc̱|6uy疮Qw;/_X6 Hvd ozĞ .,+8 v#j" =F˂n>1g, W\{\6Fdmu\~E0E%0kq mQdvMj#ED?dYD^r?{%XtkBL&6u«?ziֿ"n?} OdYfh (\t/i?č&(ZޮQ1:עn@M2&qnB<<ʔ<d?u p+RcH4$ݍ%`{UbE˖y><ឭV'? AWn. e%:o?~sLr$2ϑ8T YpV3euւ y G/NEުP't(3gQ8\zy&„~I aXa>;Lcgӫ |jW֩Žl~7#$|z\cWEՏ+q^=Xs]c7jO%彝e9 Uv%MoK=Niuj@a աJc~薳ϠRTH}PfNbKAz;E6YuQ ej>zN [e\YޅI#pGܫ0#7t;o!ܰk;%-cgRj?gEy\cqؠOZ\~A>lmQa{I:Xl<Zƙb{ s/D \.2i6p̔/[} XB c|$BfV)%v|5!며1* ,72B!JjPFWo!* ic̠hP3 q#EM KVQZ҉ Ō;dQ1C;1*1n($(!y r18 (#I # @!e>`# s:" `Ŕ!PA{&,x0ƔVBG!Eh$Z]eE kCY(V)Q{l5рU %J*6У"9~+=t&j{_"gX(WH ( ckiU#ڽv]ums] @~[GUlr >yƴطEΎf*0GGsGBįa7fQaKTpOIKao5$=x BovnsFq1U "lp6HF_jRz\Tk(`giu =rU!pr7.(4 p a)uAyrׇFi=5 s,,۵[_ϣ~R> |ΞVVe`mc}ٞJ Q0ܨ lC>d7F;ڴ/aRDyh_&UM,y1z0ErNΙ(W2;q< $FNr,p5"M0QKB(0Z%t & AEΒ53q$#ޥBcRɜRafR  c  0B)FxÓWs ( "ﶾˬ]t fG$Dވ 3!D-ǖ{B 9zC.DST*TByESV3{E9 VFˆeRH26=SֆdL31Eg((^2<#QS=! I{λw\>O!*kUj$Zޝg ?xSSϦs=|B*qYzyw?fN^gQf0SgE0EHWj0tC7O)D0 !7R J1uf: vp:> CW꿕@0! KVuY?/¨,ߍ*nQ$J\BQlۇS5_Ylv}s*njJSiѳpZY< ^YdirI'PEυڌ*ӪEhMB|og/>6 '09cs`oQ.km#9ed_샣ddx7F_%KLRmS=û8Hl1Z.R=[۹C0/{z?0FRq$Wt4 iFa~u}{_' =ysw3Z39QgM\9*AGLiƹ14aIs`U*aHe8gITO^wvӫ>t?~~۟}/?-8Hз"!@mДƛ Mk149g=^90qt:8DyVD◇/o{ǛKq_'et45j.ž\WliB:7/Ti6]E\"f8#`VLʴӾT9z޾XkmsUI^y)4-5A8ı!!,z̰)~HiD畕|;Kv. u8xJwj~w}6|᥎@^#:8z #$:!NJqpå7JR'˥0dFG4d㉅amm vg.jgW@h|yluF%!բQnkZtriZ:/a4ʟnUM2WŁ[Bw>&O2}%7iQh 5(O+R 9e&:]39u0)08f&-n:2,"Bk@%V kE"I纾2ztNz3rFVwC`<>{ o/|tWtT9;z/5kU~52S:4hBr*8ʝOD3jg>LLjyʊFXJ;4%lW8#Y Ke=K,'h 3=/K(x([;c !תkk>r(RrS_2S,,{J- vH2q:E --uq4S5[0x&( :-|N wTŬ7Y`Te+,DR2""&ZH0<)c"ҙoⳑdVm2ɗg8|tVBwΗat]{Ȼz- KAg):jC:Bǎ C(H6mXi(8m5Zڀ ̨3h刊`) NPB2%89%x%z.SJ9}v{?ƙӸL5vEg&&ݲlδئSg=vcg=m)sg~8&g3 ].hWb?&5lW޳;麃{mSAALB tFI| &uD%ԖhQjcj_a+c|]^54ɻdW(i_v]%>N,of!!- DS~[զ}vtN|oMEkaaZW mz-|>ہ%y&t6ZpO|Me&#QHY*8ܬה`dmKڔmi8/]ALrlTuhA|ڒ.^-fAOi>_+Qho?ՇgJYZh,Aʋ$T.z ao$NEոuRJ̞|4[:t0Ez[ ӏB*^".Fcg S_z_'}%7a= Kv덋7U:?-~xr;c9]M'7G)7XH>%ahp%J7bi1g;Ja8Cn ȲwΛϫ[<[wRw [hA͇3k&0\\V f'%˺:?`uBg u27n{vVgmaLI=v:..srǤ喎;M2sn}|+:C^1ŽguftkK% wLo8ְ@gI1|﷩S5"'o=<2rXG%h4(`"G5h458SzSPUgI!m2Ip>+6NQik?llcM7iF)9!}ȁ㋰r|)%@ DF?nwpqWLQ%Bٷ[[aDygJ@ |oM1? \ S?{嫭p4z^q3cooC#'jJz_ M:ϊ_C͠6.L EصoA:T$J:97. zaO,,I)﷡\RuJ?h+zhnO=\llllӲ13mg#9;CG /|O68 3ml v(`hi)1HWwGG1r)1 EuZrxDxEʝƒ(fdb$wY(@'kE}&WS弔/rGڬݾ1͐|zD^ @"bEeB `cNa@H#Mn:"cJ l*gh$keY@ V |7o̾7K &Ժ> Ƙchp)2@JF/D)E 0;;g;۬ҳ0ּz*0|Ir|F"dBǝ67&?jL*'i>u(+zB:K#*곢X6,r˩ś^}n~}{ ۉ[U|ϞDM's F/ F;+6e9gEmRsZv@J`HvHe{ԄTWD%҃H(W BCGBv] PCF[fХhre1MEĖPƽuH ÌМyRHVK|X̺KO;BD#IR!JFF eG7!(-,R)?^ @T9!iTӒSPHv5GXXkg`#LzV52 ǃ8(4bO)]ɘ7JqTQ&:Ew9lV#V_"D4Ff \Я20Ʉָ{ޯVt}BHtKZ[" Ԍ?`rwՑ eͰ!H4(KT:뒊j5Q[ow! yQr9ܟ9,Lo?qZB/ YT a7P3n.Op9Ɔq?觰۽1N1fdqr`G J]2QeThlFrMC|1/o!P$8_}D=h>DR")Oُ_UWWv8Wwz8Bx:L5y6[K-J8j=!و/ S隉?nGuIOy~Rnj la.,y;׶"u6͛G8W\>'ۢ9ҨIbli7[>jFš&;]u?~by`h2@M~S{~1-7su:nq{iIƑxHWMða$if~|D ?R0S1kAw?%aUwfqQY?tM68<]cg:Yl״3G*S{qgqw/p;ѧ!çO~Oݧ?/HQ4ev?w?u9Ǜ5l1P%W=!qXrøG+ lTn@;rNH}'iɝ0 NjQOgY`*"D?ĵ39ҌC:lOCؾ f޾XkrÎ@Z6p{yJZ#%$T1Qm$Mh6 jYLLj@h7Al6Le^Uc8yE2Iah'[T ĩv!Zěp`2Qf$ "o5E3J1&*yb spl8^ʵSS`WekMlHMQkֲ@'iqT} H}ÝeШ䜻 PxWƆTRy6ܯ/`q#;uvj=Bhvm%(nNӼmց0i3#x ם/~N= 7w^՛k_o|O#'VH+5RF-c;K.9ʃ T mgʓ¾)aml%۫|;9VkP%& !'+uzim@霁lR\AЇqHǹ ܹAŪJP9|X7tϷCK1~q#JӬv49Ţb1J&i!.x뜡DB0*@mrpVT6S`uPT8x` w:]c!=5&ȥ 1JNQ`Q*xLДD"0 $Pt'QET"S$HX4+D<b\pJ'S,_"tt6h<7^V7Yڛ|x},wy'] g.}s!YͿNpO[gn.sɳ =cpWwf))VLKR9/#HC┡T>WU!ǂAV R6%4KQ-wm]嵍nlcqC?YMqzϽj 'fo#9Y \*/Ib{Mb`"|D5t drղ/ÒzبL) <*dkȴkcI IGem {kycFSBQ&IEyPRf$eTeC#eE%xrZi-N'zūZDmU(\^7Q+ZRY$s]W??ŬX|`5>{(n!!w'^wOܸ1f}"<\-=5JEB8(=$R>_@T{`x*(8M&!qѶT[h M'RP.z*jU^zQh▅ H4>1*P$.e 빱1$au& O5cb)ÔT҇Ŧ$K(7`2c5*yU8!$yuRp[?EB@L]sc-Z ͤġ\ !UsiJ}#1C'oP gv-< T^-fasf\ϽA.?zۅMl00y,f PH7ꑮ!<"a0,`0a1i[#z<w՛qT֏ZOiƹ34q dGÑK]O!,3|a:|]h◲ޠ| ǯ@]y7ˋ_ۋo]xW} +< }Px?_Д54WMbh.ۜuia gl™-/uoAۺ^bI,ӬB N|SU]Ps*Q} Čgr>(Pp_ߴ'&E⶯^/GfOvR0i46Z1XL"VZndžH ~aOM,tNC62//cq8yMӏ:YL NH98'  7\z:u"sk4SE+2殡#ZZ@JCҤS _z% ,i:a6?06i6H(<—t:|$e(sӇ%reB0EN93XV8d)֡(~袄 -3\A`Ԃag!κhIb ;`J ~O4rr]nJ:cSZ;G6>RJjcH!|0xuղ"W>ʯ W2:e )np09sH[G xʠ:ky.k$[Ĭ)=&9ye O=cF@1gz4‚Sw,gHqϠY%:ww\~ `RPBb1-4|g=벩*Lxףar@D%+$px~ eo@ ?$O}p@fȰ_0i]*"JH.RRiU s F  *(JF4 N[-Qh\H#Df?ϳ"xw߁γRg )j"tg_M1 =00z~ww8a2L*= ]zMn2"ʽ*g If"']#~=8j q1kPFt4\M ;%,!Qd-u[xN wTV #z2Y9F$㑅H>F0+Cʘt[ŭejI>zԻ?b/ϫ3Vau|f:PY!|VꇻĆt-i{8B>BSIG ]n"L&F -sKGF ʌ ;6P 5T!2f^ig=+mF$6p=YIj)m5~EH!&%p9Pe䟽&mT9V0ڜε1O7a+Jw i'[jq^}9]iSu,~ĮYTH=i^?rK-o 8-XY$udX3Pr{0wVցٺdݹcw D@")P6us}|;CiP\:lA[r4xw=e HB9x ˽-m;8Y4\ K*wH0nNv:a|7 ="趨P¢4o|.$.wYS5il's,ME!_'J2ʹĞIY*ҭ%a+'QK|JGG ېB}B))& zKJ%bdF/cL!նdl ͒q(5UedAe 44m(|Yg7Xߎ]eO#}پnp53ZBIhTH'<1t$LEtɖ6[Ɂda iU3p \&Gf<hIB{ Ҙ68ec=# #$$:pFBb%#11)DX&܁!Nuzc :N8퐜9Lz;Gh PGN8 "T&oΊa'#e0J'uZ ]WZx۳@mcJdRe&7[]|1k*-(DIg}j}Uuh5ۨ@yVVԸ8y[8iao/][ :X-" +d-"28Nz€գ8ReG&x\o>LR5ص4h4"ͦUG\)$] 1xFnxu57s>;)oR o>qN_wY KkCXנJ_ߦ=<+fgn+87J+z)/ͼ0"RPք\`bf/ CmwϷnurp*/jri1*N*,Ea Jxc`i09sH[I^AFNqXt *4] W u_o/tl iY)eQr'kW,"!cϒ-ZyR 0]f?qvlI"> ^AUX! S~UvI COKT#.@NF}I<#~NfziPZ -q!eP[kuv|0adA].z{3?c%C&4BEgTm5FK-$:/;L. +$RΫtI1] 'u dl@=)2G>JA0z z ?a'mX+ I.\' t/T}IߞޝmjI IP<؝Jb~O߅/׫$&N\'Cp-ԿE6䒤P*l h•r :~v芵RtȘOucg-`.Qs-9˝R 3"~B iY1k3¦`$ E$DH]Bô6ՖLpRT@@ Ш'i/f$^fhVbP[bӮ,27F\rQ$l, @8`RJ!B QoTtG0ѡ_& D2UczC]9ƒK-OnEE D'+sVk!O=0 /5`;8FQMH"Z PbdKw%~̣`,Z'^D6r lNc"Wuj\KȆYuqZ?SISuYWa0X}娱84ic֢R}C(rE>{JC. (JGAE󘱸.bnJc[-tE*D$TX$A  l3hBE+ZO SbzT7[SRtX,ꈝ4&\0eWaLJ.̐b]K@\AN:MZl;Y{S4LF J8 92zܨ`5GzCRP6Z'JEBR!uJ 9GKQqOJ2"f_\l{{ZBȼT} S!_JAQh8PZMil,@@Hyh*(_CXkuڄ@<[g-aOʺ[/&ĥ("FIuQ2D$%jB@Q}o2(عa[g!m]}yi[2nyŻP ئL5WR4^#ҧmpd髒@hKQYe =;)<ޢY7nq _4 qa-Ԓ"4D iyUF>8M>);]7h:{*ȋE%YZ{[x3 nCg6⥛s"Y ԏD?^ՀXEQI9m,9xŸ́ H7V`۵7l{%&}LY%,h-dlE'r(Zsl<A7+ 餞;k4Q8FJ4؁lJ&MVDEK,6jlWߤt (j/Vkkٻ&!Sj1hm7aw^yM0:/y DEDE7@7ӌJT3 εot Fݴ{OSp(r2huT5fFj9GբFK7fSN"<6>&|IaFhQ"}_wCR^` 6%T2py;qk7ŹvzҤD˱T@=@e!%L AS'+RnC`뛶u2lz5٠*Oo (4+TC(MxrUpՎ9B~{?"wQ^0`\SLRep#5J<90]Y:@)EsOK6"\D:csjg6jI8R[i7&z@JQZ^'']&C` ZPY[gkgӮAVA%Ócf1(TBmEU|bmQqc &X)*:LZ -VF LvP(KF9)NĔ5jUjZ.C,2ceVM 2>FPCt/]XzǕ-0 ]Z#AM,?"t-+zG[V}C(7Bai8?bC߭0$! nD{L+0^ *D5*ـJ@۵@>L~v~A~}I ԛz8%`@|J Xi5+P V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+k*")XeJ%u`@6^ DެzJ C@bb%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V=_%@0׺QZ/{%Y @-@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X Jgr$8$% L s0J X{ɕ@@R .+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@G tuk2>8ŻyjG^_7o6_}εz,/gxk: KZmGsFk&.JXK_M} =e^Kf6μ!6<,٣n$+' mf}${W@j #Ff!Fih$` t7Nz(e2ݭTgY~g',V>[yIW1yљK8jVv" 2[_Q)8/_]?qXb|ogh8;|^g-;sR U/6Ͻ3-?mE,m&McxWP9 ʯE34dU45tËe}\ʏ@SD*6dMwUR(0$"JG2i˳mҥl B: B,m JVDUnH"HOO{2zRFsOk,rխ{/WJr\aR^Ks%+2}Zbyo&87 \k2\le[r/_t:z!J<4rfσU6cMu8փ;gi65).bJ!Ke.! +|^$W,(d4M`b){5KҩnXdhԣcwX}}yҠ{AvێuNAT p!PϾ'J]%lBQh[Lt%Rʔl5 JR4Q, ^ţ\t~bjǀ\x2l(rxNO*KX>"Bp0knu;V.-ԯ&kgwHmwmwmvvGYéYqlw! Nx}(@xXmOZ9.O\ys 7~ vgxɨw>!GK;.g_>ֱ݄Tx~qݯ]9:F?-O1Ou񳍍-<. k;`J/i_rKiKKK.wmI_!‌s6Agb Op(P$!)E+F(l"X+`B.b2`Rqțef`N !!!xQʼـ{E }NrK}p 1Ym;s$䌞:| LO/z:ugKL:go!S{NА6F::[K:nUp ܲʘgh՞2#,Ҕal)Ioa;fzfN]T{K{Vd2Kpl88iGI,Wg/V{+bNQ5cGbq3K?,wAc0I|u3GU ( E0-(:Qвc{㉲GqHp̂ 0饷N8XFC@L /t!H&<߲x.4+ pNjehԟ<epWz|tHjߘnd,#?ծ`3*oXTY -Bc0k7\2{fŠGrǎu{/x>f)BǞUD[<&( d)Po%v, r>:/d$\t߄y$LViǬQFYA@`ڭGgQrly2['gYyY5 0XJmDž;yyDhu2.[^b[0*9G"8hca5"M0QKB(0,H92J8)@h|nܖǒBmMBrQм@@itL*"r04 }$" NH&6ԓSZP)FR8Qyex4tº ,DP /(Ftg I rl'fWDBbvPiuJj$P^;+p4PD(aEm$8 Vx-kF 6u08emH`))҂Px" Hr1xF<+b(CFg}YpL'rLEEeJi7ƅx[=gߔ<i`ٷ-ržJt<5?.Û z?+o!}F[Ow(b{KD2}; @ژB#x!]HA)̔er , tQ?k N[lB/r@RjXs{T/>U'}d|u*LUݘJT%Bo<?*0RIHQgnRdSaCH[`>&{ cK#Ym>}z<0{0gl֙k?tn~۹Mp4vR2xo60|jkI֖_u[3ڛ_,@h}*3f1YppLtf?Y38qkod[cM~f #42|:Wo^R FӓRU*`McZY8}sKXuw?)çO?OݧIpO[~vt@]Q{MSihi*E\u=s“a->:VD/_~Kaߍo'8=Xi&|6r칺W]fs ]J(1Ĕ3=Ҏ W NZn ]ҷ*H,/ǼLE BJ 6qlHD(Wob#Imht^QSq8o0ɻB^`[XŴqzD9N~q(C,:e|[G?u!ۚXh@[n3,v>:$L߭]vNkinRZV8KhA_~J]yk^M@a|,ܢA&rvziyYϷsZZ(Otگեj6xnf/|NX `#QHYj;x0k5f rDkB{l5ͭ'Vjkgi;*(_!P3I>tc"Bem<ktn+1#AH pAhWpO0CQ<,@(vGj$r~]`FʩR ˽-ꝩllp1^2dW80?zm<.^z۝p5i8W-SX459bڹXoU6Ԕ6|Ge[<~b|[xJ8Qi%LJNl( ^9ADZzSDOz$|t8ofYd>ZRTU֥Gm7 ,3AI֌ٺY3G)GхVsuʬ U g=Jcܦ̖|WӇU`|}I'Z0,5vF=L̀lRk %IQ!`V$9^F`bHHLrp͖M:=0Q \HP&Ga $ >r0 !a/;dɘV!!@: $H[XA`h NȺK9rXˁ@'iuOI{%Y]ITVwO8CGVGMH":w#F޳;e{Y+"E`D Ģ{Ev"O4Z$H〈$X#%TR X/wXXӔ9e;Ory q"SÂӪ΄Nk3 o_u>Tɵ]+G*T>ōI@h}թ+UA7?kOd (]?.W9,7T:+!Q߮ͪB(Y@fejM oΑEs%ׯ.lcfou u*?sQ7[\]-)8/]ݣPuH\*yVn>I9YI\>3%,``B Ed[Q[e.8=?ERKY*1{aԸO.}ohG#|{WG^$.R,>?a^aV| lJ7LraUJX/ڢ=vZڕ^Ynxz(0sNnNRnd1EL<iX{)) JnHb΂wDðq$$ Ȳwwz[+⷏Z66+т[{v jx2]x\\V f'%fIh\Y:8PME D S˒z0;ɔnԓLjnܷ;oL6K2]z\<}>:-wJZni tY.?טγ0V垻P# 23yj[m3{Rָ\LV>n[L*2~{+.30ǁҏk'/T3֩9Iklh췅X,m*Uv'L'F<퐕H8d~DaP$@TUIOzG^`zho1E`󧔅!xnh6942¦U'\cHi"ӬK}s1)au:3XWaR| ~ &uDhQhc dw~k/fgxn3KO^fzF7 %elJa^iGjd~sVk/SJx@Uܵ X-`L*3r Lo_xz)s54҄ {( g>T[6Lr˷^cx5} U&/QzLN';s0 XpVvH8H#9!SOS=$ >p:[kj/|W| /aTw'M6T_h|[)_G W[w8 LKf%\T=BR+`J*Н|_}} N N`:,(* .bS{:ɳitCo+֛V~[٭iyQ^Y,Lnr|@wܛ܎i#]Ncw?Ϻz3݁^j^ F |\* vJc(Rc6(hsxBn6?vC<1xdTCAEA!"G5h458#joSڥʎܝ-nh,)S vP7G7+~Wl)>F"7[v,8vdHd}DQ:tԡuCG:z2K9_iq7B -#"ᷠQ bRb* cHsʝƒ(fdb$w>Y|BA' kEݼor^ 1MIeڦ R )ґ-V[&pP Y ׹ ;4fM'YFoԣHC|fbX |#8Gp.Gc]oGW|Y.#!8:MY5~JBzm=>C`$ytd!Kia:˸xd/lz{Zg 8G]V"ȸ"v)DBB"<zԴj/]EZflGEQ%|Ćb\1ǽ ip:cHIGL t4D{S(y̝vX-"ro0CJf/h{F&Y+|H7mvuBYNx{uREN‡r+HW'OX05]1=^^xʸKX]MֱT~c F *X䶰rd&H{J+Z(nE?6Cfvٔ|$[h~{UZ_.0`6˯i]'Pű# MKq&rb3K0(F1ZP!03hR`%,Y/o $d:bGlݽun,dY(qvҏz4dT/x`fy`:BL Dktz{PGqHp̂ 0e:3N8"$5ć@^JBLƼ)P&x9H Ȣ` vbrX+?B"s g#g= ],io~LkűʲH^<ש?/{nb|s?S Tϭ%y gQg{.،)+8`VB9n@2䊈A1|!C-!3bsK|tG;V9JVm`K0B!GtJ@1y, Oy%#a劤c&S&SNjvek KR5,3JFΆ{,s>?<ׯL?r( , .ms9~Udݧ[nJu7λdWDF9@ A SAh*XʜB"(ȺsI`^ ݔ.: DFǤ"J?04 c  B`NH&۩ASbơdMt fG$ވ =!D-ǖ{B`fy&ҳ^ؿ_jihw]8EI% |y(07'a1^KèMm!i8(U E kQ@[3^9B5xm;VUM.*kHZ^l0qQ )L'|:TyR(Lix?4coc4j B* `f8"apϐL2%g'X!S`/\ )(ԝ Wvt{R@^.{O%P0 *:mqkX˥I [;o' Y"ca> }tTJjQ7_IKfP* )>m lhã ѦX'.a;0aLMA<U{}]u㇛es"0{@ C36'|hxqY-ws$'_nfԷ~PHw#]u C:{EfyBxR'4iGDwc'0գ5j糂1t'e9_'S)`Im<;ATOl8o/z㻟??ޝsX{~/0ҙ.~6 ?3_6\CxT 2S7~\ <c܇4[C7_r+/?ҥ$Hmj'm8ݭ`M#@6U]eX\EE_SقsWR!|H1\tj*U]>NezX:,>zN&|郏HrKh!XĀ "D& e1Þ2X荤6:!z^ 6,-7u򚀥uYL'{$bH' 7\z7ulavsL̪TGC6Xr4 ˛vkkx:oZZ{U}ҞVJ)]xTJz)U"ޗR}T ^RTrYqXh/3]z sZD)Z0la% , $/UCt bE0.P DVj3g]$h1̝BJ8F3XkC:ŲeD#|μ^QrtL%Ӄ(`oRKޣAX(Y}RQBqXt)np0sH[]D‰uXTTuI>17!Ej.Sl3f4j|sWL#,8ZI}>;9H]{MZn0]rua/ 3/JfoPoOPQMhv74.vmE}O$qD=-%6By[Q^kypl3_PAӸQ1\\$x/?c%f.5%ALgD2}kn#k]Q> o]-vuڃٶ[ 7?T 8v~?X4Λ []M/7)7;+|28r_@3n*2ULa"huEphxrڽ\\VG$^0',^B;tς ʤKX&*^^^:Xz+DzuD=iI v=0̴qIFE*qz-wHZ tmsfEH(oYώuf1M|,[+(j{frMmv-,'j$7v8nϭvLPj={ ̈́1E5"-#sbv9nx2J u"":׶諙v Ë &e{ʑq؝frd<<3H8DaP$@TUIOzG^`t[;S<{&g?|3xc|3ˡn7Dja몃uRH!CP=6NNU]y:?Q%n$G1/j:Rj f(H ʱIeFSf|cC!fH _1zl2ϙVohzpC,=Ғ`}HfO*oF>Zq:eԑ \`|$BfV)%vܣnT~ ܆6;ց7-^-\/۝37}3`t1fGGshFd>ZRvBP=ǥoW12#XP1g&ʭ^3F){х8c[]2B"ϣHg7Ɂܱ^߬֗+oK7ݓ_7(h4<o\cK䔃KI!jE*[+p v rC bc$FӐ? ZK(Ij ") es) Ӂi]$U$YdCٸc[3kmkPU iU30^aA8`t!F=r`iLgQzelYƴ  QzHLduɉ`!LR#1s)|ؾ%^v)iG-a;5)ZkAGU1+}ҞR8XIr2D%, "^HuaqLâJz9\&DD$(IDaQ[B$Z{3JŲzRwBgrsHf˂ӪτNevM*yP+I3e%kCUyd[́Zup6K9rnWMr5?L'wgi~hsIZwKm{YT2p-@.*@Ye`_Yev,3wb6\Bu$.$S]gt{LBrw ^O|ZocUvwu6T}Mnԝ5w~jj)|fv3=oyPIO4ފlNtNt tgmҊ#RW`F]cQWZ]]%*+TWŏ ]%5?uj$E]%j8tu=zq^[K0Xv4\T[[03N'?x`:M=&X /0d>@8C58_>7>AK5$_VO"wQEM?ZB@OPӸW۾z5GU"XQW@7|U^]BuEJQW3+ٻ޶r$WÉy)!t &=`wW[,-ىţ%YG#Kt;Ey+VUcWJE\Dqe?oU_~z3aΦ0*ZOY_!޼ ^QUN1ZeLh|yj>wR9>`G6*w4 i iS!Jͨ"jL0~}<ezFI"_uA+8Z(8IG:%tyi%݁Xrʗ%|)_rʗ%|)CACAiqʁ v:~LSdj9v? R9R4/OAiK rټXU(/JR(/JR( KR(/,,/JR(/{pVުjvjyD"[JDFڕD,EJyR ')IJyq(毦$%b')IJyR')IJyR')IJyRk\mMrjoǧX/RWGgrc()7z(< [zكP]"t.0~w-[b1q]6 &eF2 ,Z#ՁHOy->̚ZC I^֫W_gҜM{+/|w;=Fɲ U!r?PL 82DQ1xΒj=k7y/Y~ӓ0pp[[ -l񸻵͚ fRۆoi@xpY+YYy8`.Y :;0 (15WNJ#I[jjs=Nx}s"[mېqӓF"I-..fBQcaȲP!Ě"^완1Hbc<)\]wo'[gNI4a6q fr GMRc6Dg6vdiWU۬3mh'EVtMݦ9Ңamq҂oGDg$#LSMr$\EOɫ\sTP:ZZJ7s`r6 L 8:t1ƶ d 7U^Vup^P3%[MڎgTB;*)+!Gbʊ*c\6 "*NYߥ\VA~|癵\$&jOMҒGʣNjT\ D2)8pKd& #]:n:}U<09, c؍ڡ7S'7NQ')5KQzZ-9ccюq:H0(8HTcfjcf**1_`q\Hތ=l?] \LP9v&mhy8)W4"[rRJQ.$yD "q&Ja.72 sX*/_<|#p-M3ἔg}6w|,PK4uNL&QRIO-W@&w-v)B-c=䅚е}\άTj$xFW6άP ꝫNBJ$T^kfE4 (v.v-lJO3u֭꩷g$0SHñ&)g!r*gjmq;2܁{.vz,CEwF._QXt\9r~ǦW?~7b2"=-]:{|06t:R3GTmtAZ@*2UZ⡐=Gwg߿ۨAҌX_圚WXdٜ/ xY6ͱejr+Zw:!Q Q{F/-P# Zi#Dժ$LZ{-v@7hdqq,<҂F0e[kl{۲djBgvB=N8)u>h}QJLWEEh_qήJkBP LΐJ{ȹ4%jҺ?U-ۣJg޴'vV)j0Ҷte\,k AjQZ)al4T$ Q8! [նt r (Ip;m<8Q\D"(tA!Zֶ[#g\3>?{&_wpp) V\fVT~Y|!,F;rDZ+J)+h4t4M::^JYh b0TIp9^*߻ʵ̮\ϔ !ϓ/hFeEAکBDD@X x*k\n䔗A偰$3{3gR̨NpeE61V ?ϵ5m$&J+C:$ $#')oTs0p(9OIR@&-GڹiЂrsHN$%U$_jˀ(r0!6O̱&#_W#ӛ/qߔ#zF3\[z6ѥLx/Ĝatc/3gVyg1x0C; c/'dHf0ӁT!SQ/"ϕr^G(/_Ǟ&ϦkI䍣bPWkf;\<9sP2Om`C +ҝ}w+[ ENfͽg|QE32y7jCN05*pj5u >{5*ύYӤug<U_nF+fBKOxm_9'Q?~3;#oh .aH<<|DІ W0btcv/56Qlqh czO3#y`] cg:Yl״3w*S%@6<%AYCET1Qm$Mh6 jYLLj@h7A,Fs{2=GT`Z::8NA C 2$8I^LBB/:UxbUG;G4( KKΡK+t%A)5|ݫQVz}qBSP‚7bZ ,qU$O'=q`J.s%f1ϝ<[bi+E)sH2Ƽɱ䒣R}PTyv04sC}5pɃ0y"[ІDU*TN8S%&`̩6cohdMN!9 e3BDLD$4BhS()vLq?@r{wW.c}|5W Ϗ%dJlU@Դ D$uݽk`_[ :UH A pԣgF*JJjAB l)-=쵂X`-#[B]T>E|X!EO"rr*q-L"~1 erR<$2r2*٩L-cWJIzJQu+ ǻm0*~zթ~;ֶo^Moo9w:AVm_>rpx? v>Fr>|}iK޾B㱦{v+MmCʴom\}u[%`/@˨*'WA\RT/::/L=;W{Uop~Uf>_B@]w}v\.B )@9 eKImoL٧u=kv=t Vׇ2kʦr~1?*`dYMQ*H9Ky 40b hf Jq&JIaQ娀r$&Mc=9rw LSb&bZu{9+¦U'R(ICҏx*SOľrY%JKD*K`f7P7bnr &i眨8sN;'<0"ĸYd@O#|Q![QzI8o}G*T tQ&KfHڰ( c\̂* ,X"3(j5-!LHLdO19xHi!٤Wl|TQ;hG[h}Ս ;ݩ[T[-#:nFMdV?i$63 }R{2YφVMCK8(a AݜU#G9,:(JGc+Jf6kkc\I ;ebF/K Τs.ugf܎R Isu m yws6nIS==7'7-K(Hx䔑+2`!7F 1-.њ2MC)齐MMuQM&sY0!'ckW՝ۏqŸ<]M:Nkwv+IIM.$D&zC Ey\r I{>JUB&dȁ !]" NIG Ǭ" dT'jOVn}ۢOE#VgjDj:]k -gL 5)6\F"N  |^ mgVZN~ ;ؑ:KmUNY՝6s{;#xO)LZԁ,NP!bedTYbCƀjD L92D]Xǚڱ(k*3 yb,d#2ʘbPN[9,gg1…MR9o,T,0ZTvuîNGw]wrkfUJW䂺AB.o{!E}rw3bu lGuqvcM'\B+wnhy~xzmyѥṕKDs1ŭ^z}37M՗v(&{kulsPZ+uHu6 k-<4{pJXV^Ɨ_hL6FT :k+D JFry_Rd!,{^<4AbcZ&.͚a:[+@e;#'-0+/?uSUzكB:% j [csarAkGt .y?s5ugحظ9vܑ9MH }(O>WG)+7iGK͙0SB;pԁィ1\j`8Le)tJIHK0O PʋL߸oi[҉R@QjՋK$Z7_6O_%Ht"r,E0P܃6:p' Y\IU!IYu❣7bG ZWoyߵ2mଡ5+ |7Go,G<ݎC"Vt|3&ux"$ Hh©윷Qz&Ki@}0sWygrUv~K9Pe [iPkCV㑒O _HaX\h2HK†F *:t$:';PγRq^$d炠8'gU CdڢcWUf'0r[99UmBRE>}`V;p`)j+ުug^>%38MojeRrBfM6o 32ϲ]yk~9JI;N&uqUoH1X1e@u޺5V/ D!eɭe Jj@$g b4H20Zw5{[7'J;͸vw"Zg qӏ{V_^OW $l_1`Į U<愤ڗ8*d[ &堲ޡNȹFtB$82Tpad fqcPAKcvUjٽYo $ v}u2)fϛβp- 9)XS` `)L]h3K("Ȥ `ӼBwSjYܸ]oiko*鎃|%MrF ~EJ̻ޯ #}Ro*w WA / R{?3/d[!Y|_R.%C8>qq.p&tw))4f~m,9 =w]+h$s;~0/HPNK9׿-#S9?Z/t׻qih??/~[-}XRu{EM.hh\4C8}}c!Qelo1n/.FucvӏޯG3Kx?+Np4v&'[N mjɪ-zM͈hUamLQh,]|4-&zrf?28=s>٪`[] rSJX#aE8GNRo{Em\ N:=ףd⟳˽пH'RG~|~,ÿן>/hh4 _܅PijoѴG|w+7xXv[c[n/I(EO9p5g <ў܋Pi_Ϫ'0D۫>Š ױU!VVb&K=P@xl'c9۟싵dK*)3(|FWh4ܡ֐9k0C3M;9(&Hކk=A3u4(ꐋ8G霳"$K&4( r);ԩr3Iۛƪ:h&kai oڝZZ QK]U]ґ◴!+^ bLH9T37!ޯ#=czƣHʊ>4jh Ipm,_0ndIZgCV\HҨS'I+#IIP$iiƞ4n@Ꜷf (u# 8hX^̦"&W#.zpJ;{xpBC"칌%Em#Ff*|;)e<3c㐶djmB])=5ˢ MPdKd4H v6`R3U P\OR9U7Cn 0ld ICCo"``/Y@!mK%:m R)1!:'1BLs q$52ǎ6% l&ݼ'0:֗rY~*i[%-cޑ?dPҴQȬ ̬Rݠ){. 2>o ޭ ]g*޾@outQd`0>3ͶCZGBZwiNYR{ܛa/9P3>/zfЮyh*pd\(ǵ'_4Y~+ ~Ágzȑ_m^0'3f l`jD=?ŒdI3jUXT )cΜ('߰v{O{4_Q߁n{}"SͻE>"jG}2%-\?Wy[Uf{>PE{ۇ>S05A:{quF 1 VP"2dAQ/-!Hɧ:Σ $Aq#rYJֶ %,_Jo+o|Ng]S1HL4 P )K$G]69k5I|u)oDP-8=&pE0cԘh"Ĝ=:EEHQLДD"0 $P3Q.T"S$HX4eF1.@)E)_Y,G#4jGtc[퓃W>)釓c~!n~ o+IY%ώ.gvյ|kol7[R*=T/V`(@)9?=)4h/Lɨ@B ] 2$OZs$kR8eUHƱ`c)U1H*TM(JKb얌QYdaq,ԅedQeႢn||ڱlwbfniw_3z?ĶȖ&0%@1P b2TȜ`%uN-~ r#(,AV'iEceɥP#$&GDNJe*EnaIl.;ڦ6ּŜ1fEM:TWJ8"2.SRzQ؜8D0 wZP@ !Ԣ= eM,8!p9(.$G:*[*p1rvZF`lD,?:]KhJTjki(L򠀥,\p^CQRM^Qŭ=t }eۛ6wqZ9MsbxokW_յv1ß5k^kG2.VɴOǘJȠJ 7*baMucGSoc e_cDBK^DC#aQC|R,bBV'_JRZbL̘:hOg&EnVmG+XuΎ]N.ʟJu^f '=@22kα+l]n]M[os8͎}}6e-6z~jY=z^j}?7=;^aY;:6N ,Z(9y}klTmN8?ߎkkך[I7*}4/"sΛe*r4/Rvg/HϺ4/?)p |\ G7J$rxʸ(+d%7KU pTc?Z?wyk䬯P4Xٳ//!C>&.85gЧA(onʀ9$e) bTmAL&ʌdXGhk1RM߽ϷϠvۭlVͩs֮Mߊ |uSi#,ɲo:i%PKH!c 5A)gRX\rT@ dZJM>;Ygft^caf{]CZbP3E>`K ?7s5w|wǙ8c̎1ǘkuǡ |uyfLdx^^=Wc'zёZ6n@Y-!2O=\:)&)f$an|-6ur 9†gt|t7g;;>w.z7oQ6G&il;m\rp^tm3_t8{λx) LD tҡClgh7eZmP)5nǼ 'l_;{up41!LtF249-'*HU$9 Փ yN JGkTǣ 3cO}[[9¦˪Vi 9&ut -)@qW5{Z5j>Z{KΓ*pm+<leM͟%( 5;5ͮH_\mO΋hi85eFNQL +td@pJp^Huģe  1DqŜi80JRQk"tD,hac`0-F\Dn/2Γ>-3%}opOLEobK|~z1kC[Y ~<> 51DM+YeS<5"Em HpՀ;m<8<9\DBh24CbQS)Ia{pAU" QWi@(29q,$@M|i\6!Tl=|j&BiNScN3m:W@4*!M"Pl-v2Py (b}h_ 倏xE<6>;w^-a_V-Ơxs %RQS8_\JJI*` ʊ*cP93-u86;gqo_fيD8,i#Q'4M\&dRpV1cQ4Jf0ҙrݲňkM'6DB-?ygLD` ur+]9{ 6e%yRwcc<8}V5—T쒶=v<;tt&tPu4t(> i*M$hy8)):pԁ󂣔@PQ"8 FdEL.rLPC)eCV e(ۗPT}uŴk97:?5=5O_JrcYh$ LG5ZP+#pG& 1o条oDM44w.79/w- v+8+(vEـo;L зCbT Bo]!Aj208;W%Y A$T^kfE4 9,8Wt/aK8c[NIH[EBBޣYQSHJ jSEw}5 xB-] /7q؛`|:~Ab s6Nr-woy~6? bk26^Eo.^=>y bC-]OjZ:FW֣TgTBP o@;\Jk{L4] ]\bx2zo+˰D5of*ة'd<&!E1vJq0;uEռ}/-P# Zi#DX$LZ{-v@7hdqq,<҂F0UA1rIN^6]î &so}V-?!rѴ\l'wR#]=9,v7U3UAL^IV#@FL9ZX(YEBP LΐJ{Q(QL*t˓UvAT /SOMwHF( }PLSk Ajp 06* |]"rLnf__fψB4mM̭?Ltny%#Qox7aRA4EH 6t4 kFa7dw|̳Oq>1ɖ6Q/4j㻢2|<}z7qݥ^YJlT}Y8}܎Aqt~9ǫ_~WՇ_~/8賗g ʼM_3φ_Cq\C Ug{\<^a죧q2#f ?~}a?'o~ө9*^ l~;c{~EM_ש2$mEfEx!GA}&&n~$+;y5Pb JXFVLk^%rֱ$IO5Ty̩H呒68BZ\2:-c;K.9+kn#G>LotlG˼LHZӤ'OxC$(v=X@0 L|H|G"eV2 e.k(DzyB!sΐ5,U8 .7LT4V9%ﹱ6>ٗ?BzÉx5_^;s#-}h98#i7&Ɂƪ#(ܤR*5Ad`"6fN坩H@㱤9Q2eQ'Y:2 Is LZI笑9u 42Ьqz^kg̀ 1fpLmfz'zJY'?N6]+דOA3+`MzTqsG/35٘w WW^W?{O撤S'HjpɉT2h0.FJN74|)EW}NAt2'SY] k]W1p%RGV1pq 2H́ 2 CK 2J}"%.=_?5N.yi'^N6Ov0{v͓+2wwV^+}Cl qscr: zDKS.C>Rro$sNz Bڣ_zq_زz-&[;:/'=DuΡC; KA6,I1,>J:UA`Ɖ#5)EvU&%巄^DRIkR0SM*7v&n;H5njd/ntùHZ܏Z5dTW^f'Nv9 dzrg/ܿ.VyVE%8(4#F-t$ BLa #[Y[S!·MYrQHR(٤Hs>m2nF)'fƾVGsʳ?]Ellg?d˃?i~{ok=K(d̎+]X~4h!7F3-9Iiey22!E1R!b,\2=f] 9[;W8-v҂sy.VjVuVCύK!iJr-ܞ$!2\0Rdc,})P=J-Y|A2@^th$$gQ DK+C*HN5_x5q֩/EcF a<Xm~iku0U-1m|cT p \LH!Z@WEC,-b5q[ĕ¢\6K]]YEo]A)IQ/5~n/]y봛6Y^]6򉲑g*DML5kuSH)|&u:F3ea9LوdqתgR!!dPEP=ZzډyURܷSU{2*{PEGlKq6w]^?$`ճuC/Ip[:Fol; ńvv~sO"BPovw3U32wJ&XRi[:2>s0So(>ǟ5ݵLRf[:%I\s| V͇m?[k5jGVi@quc`TuWOݜ~Nh (oÛbW*;j!Rm,gt!!޸)|边A>ݩٿ L.z3k6}Z.6T >޶EfY_eLv~l—88sT[M7g7@:3K<FqB*I%`RdGQCo)ȪgGz[~voJwkբí%*Ѭlx<6q磗-4S)gi>buJJt⫹0һ^5nUKjm| 8^]Jm Wtz.4磳rdVB//fB&Yci̲<Ј߀ci~8cq{KrʹaG~X4mxdk5S9ΨM=cn2Lif`YiPk-qv8bΉR׀Of,go|;-u֞RNֻ{SSI}5 .ں8vTst4vGL֥HK.(7E)O; ڟ&;_ʟ.~X1a./C-լFxp|ä]GvFt=#z'h%Ld[z#VnI9N| ]``.Ot,FZs{/i2[vKD{2aaƨE(GKAk~\j?ͻM:7k[a5ۛn?}G_/?&ydz^ڮ Wl= FsQFGˣտ~<3\xy}-T0+ p4reL J/rr.RyΨTB'T1:H (à "KO]jn\|˹q^~sOo킍/DJEY4K Bё{i)$?2 f ;孢SgLqF_lvGokc=b2/.sp R7 hoxR4DhGw{VdN!4Yh@&Z+B'9Q߃zQz5d. %nsÏ9;t8Kշ{ OAOpлveuL?a{r^o4ȭϮ&{wE<^w\;l>?ҍ}#G+;P DJ-,K@jRAA*r씫ZGFv|?znMyP<.K cSJ4L"BbVZeء̋;d#;SFiAY*i@K;YcyL 3Mje/\F p_*-:yR%B*Ȭ#ۉI\9VM][|N7_ B-<2_0v&'7Qyo)} Dpؗ.JZ(匵bg%l,QҸc*4(v<gODVaϜž[bxuluve4 ).j< c h5$mY$1yt\e`hNŔcA="]AzHh$Vj2,1-=G&t44#ӤFo˥[{wsj^~ݭަko*'% ̿5YR!W- 1;n~u䆑fsrp1lO#9F/y'Q H׸` 9%080;aOZ "x7>#xK[O~ŒCK%X/u IM7k$=A7^p{X8( `N#SQ&J9]T'Q"ʹ[LݽO?BsPwt/.9{)f|hsZ'hHfTG}P-izw5m}r]sE2@"Α9v_<:VK.h9߱G7' _"c ʛ!WQl5Ef'i8jQPbF%q)P9l+9C2=.i&#$mtʣǩ媨A5lw5Cx*O!utf7?7kp+Zq~! PMBA۵Q} Ck}?hm&x3=@K樧]u1V6l?̗F8>}5IYfl8]=ThM+6$? V] DZQ!\ LFxl*&ݿN?Z0"lA-vFP; *ȺFT`LnSx?# D7aꜤ6t^݈ϷM8lx[ Ţglu%p\TYsR0J sI1ieSgjٹ~;^gkǨy9ye;d(``<Yf#kQ=Ws@iOԂ5$sB# 0\rrtЏQa%cmo!V %\*άIAeQEQZHۈ1L -%FktXq?@quWIKyl ^)r^)E%d[,(q3M/ܚ]+n:a+6nznOX7]j~cJ+GGkX͗\RY̬QW/it\ǩzkMZuG! ;r&_Uzer42K<`<+ǼϹAGET 5ʌ) 0>&GyrF$0sٖ.ZCN 5-5_/6ibvh윜L>fw7-7QhW;?;ܳPu)7!^ۧ,ϱ0lMQ}Oٞ"[ E]ltuEXj UL+BuUЕb[UVMWWՕ5j*C߿I__w̹p˰^mK5k/|ƺyAr"F%]{Fϡ>DQK/,34',`ފ.\ip$8 Ӥ6۸Wg.]E$aXfz0<9* F$b$Lo+SG]rijh+vz,Yoo;왼3e <&ݧ!sfayYXJ g'܂CTB<5Xrts~f-Wuoo Yx128' ڨ" ٨h%YCd%,k6fYFciF 6-2dHݺv&3TFh%xM5u x]^G4,3lQ !lۚs >l5lz Rt ?c 2!t 𺌂A k5u x]^kYDxy5u x]l mĜXG2֓ul,Be שr-ŏ]i6{ ~hFzshW)?9X=eGVc7"p8FaTnKVK2KYj sU+)4$H(J=;'5v-ز:cƘs yCjZZc#N/ױz "O7?\2>By]eVOL"ruHoeTIY$/0( O@dΨ`A& _u1\WyPP189lcI5X4IDr$I;IQnr+ CQ ksTNGA'Kzge gU؛qҝ)"]Sh::e/9QV62? /Zh˴Q&X^I0A J3G42.Z:AjWV6)S1H*RbPdb#:VjǑiRs\h7|wT Qy>F m:p3(q](v;_`gʐ n9SVWGni&?j*w A1sʟqAuő,q9=0hBx>i%sq:q/ l=6mKr\`,yhbԁ$5݌%/np8yX8( `N#SQ&J9]T'Q"ʹ>BVՀ 9r>u=@6C;M:qDG:=4&A4bMɕwͅWgG k 8G}hB[a/i$&)joNADVA2r%$7C:^j0̲OBp'`bYGŌK-S+rv; WA]9֒8:.i&#$mtʣǩ媨A5t5Cx*O!utf7?7kp+Zq~! PW(֡5o>@hansS .O C^+p=FKn#@BTοq(.F'%c)N椖r*[A&WqBx):L`U,.(pbAوjـdboN܊$ݛGLC8I^Vg d]N*e0VA&I)se<ҟ@0uNSGf:n&Flyng{m$v^"o 4aam4by YcԐF/dIQIT,fUe<qŠ[$cbTǔQ:gaȰqBZA2zM&:󲸽yTsx٘h{ov~[I[9;e jMGxt%pOlu}ua~; );jP.bT m2 R#K׏/V?bƢX[btBUTHQXIl#86Hj2),[zݜGUueǃͻ9?Fq/Y,A>D-3J0_T(E]tt萼TN @KiΘ8fifp""S2DNQйd\PSd) B'R}+&!F)t`)֐G/C)O9aAH\eY"E Ee3f淐8Y̾?.^MpV_n{t~kE)_򩫝ݹ_22ҫUxKOo5M܇tunNnڼ} V"'KT90E￶_d㢛6;ٯ_[vܲ[/~wow_|%Ǖ>4go\~}tB~Sz\~tDžW&ʊO=]o٣ zwIɀuт *مXl%p&VuRA-uwv}NżXϤoկ0m>ފ"RHՀ4϶xq蓑VYG%_G2#8MAEnȥbT]i p[P/w3W.γ4T/6}-YGܣ\}|箦Cͮ˻+_yEJ`W60Ab (g#Al!PcAψġ]z)jm]\F| j%*8,\Xӯ҅7(- QQc!bd) [d?x3IzN+ω"yNӴ$ȻE;n{;]LkV ܥU:u^Gb& /*(kgO;m:R;тP~zr륔.o_x<:j ]l@, U.Ő$cQ(-e"X#OlglkC y˸e7PbUW^z)dB=%ĦJZF/܈zMPDYɹ5ɹ~tJóǪWR]gY^' `dyX쇚Ż/i: ϵnY6ݯ|/8[GqŇ|ξ>ÿNy~}~;+뮷/|>Nq{Ny^/IJyW@Y ^y/I4w?/әid>P_ `,#!{lsn\<ߖyXN{s&p T +v_?_<h\,/o׷W_M?MKϿb骸]V?/C o?]h:mz3~UK^de2`.־?Vc{Ro=?нO[RM5E7f|ݱ-m&睅70Q K$&+ lnZ@j`iyRPH3.EFrAd*y8 ۞ m49:'zj잢r<^BF96'{g#qNbɯTFKToS8Nfp.@2d@fDeSF”Po;ʙb />$6Еl($ج;q;K9 6bo~£'lB-<1"Q]bc_ǿbKzjs6EvB䍗јU[[ŋC14؊d[1iqNb) D>$r)6l9b.HAf3"{ rN33,ubLA{BHgX]Cf C"LrW`&M;#tkbzl=bdSq>& DrUdwij>L9Ϋϫ~`t4F4NJfnk@+,(&ftQ3&$Y/AjS9sج:2S̵'92b[-4O VMWe gȅs*M.0VE[8LqZ19*wP9.g}g˻gF"ud-yP\ dSe* 6kU$c;{TN|HTȶO[tJfm‘&j , 5ۭݩvwEPǑq|vtug!j|h@3)Ny;@Qugl.KPJ^D( kK.X%”J1QT)J4*"aLH,b@ -YH2OO9넦ֺͺ=D7PVz : JP< z|ωFQ^ U`M&tl 9]@$hbH+~dCoJ|uwkbKҦJK8kYא U&HgAHktiʣnGPB I>sߗ?ݬ{Ew:XGw"4k+|g7g׃˧餛p<  l|dbLLB$I׽HzhywRֺ`U"bghQF4ސĢJNFȢQ #Whm:[ ڳ}\ %UE,lDh l֝=#fSw3|ψ{{fw?Z{Bw4Svafzхl_G!u _PIiodRlK15HFK P-AJ2h@_I1P,>)OsJd5$ZK4(zatACsv~ѥ{?nS,·ѲꍼVf~=|?Z]oy}\#Cֳ)/<ٻ߶$e -gY]عC ;'tX㘱&Ŗ$J&ȉq+C%2>99 #  dN 2u%92cH &Y&uL ``tu7ܛ*&b(ku|#I8sgao*n.Dc~xIb®2oHfNz}ɏ؛ٳJ1#i֛pݿHS+d@,q9'#nI()&8"љRrxhukhu\2p.#aE/Q,U_7̽ӓYFmTӚ`WDpt}_~?Oi=}_oO߿{C;Οh&˭u]0]11sY}MK47oZZءiejz6uKSak\||mUfq$ޏ׿iy8횤 *p=H4 urz⬊Wy+K/S !V|(1T&ل/'q[}6e7x x$Z^`i 3NLj^`$)I_0Ǽ|#Hs7KNy!qNZ 9gEIZ-( r%y}Pʡ'Wښ+'?cvw^ϫ:_ZQ CwZIx/3hpX FQi.k a{͓׼~kFzZ/|N/@&2,\},%gJh}IT&]#I_DCU U:*i:Z1,7^0!Em%˥όtu"0D9㚄"*YSB/"Pt` {>>5FlQ< X,@ \hPjd/5ǫ3n[s;$׷95I{6,x|Xٖt8/ՕxbeD1 JX qP,%|BLbJM)!1runkkrJ9V)kkS\`$J]c974sgX;;R }VBabGelb<^ݥ68¦i_\ ?]Ǔo=K(fRe2fǕ.kB =-#<#AVfi('!) lJm Xd2{̺,!8#s2Zَa8HK塠vձ/jʨ:F?p)q3-PIE`$F& H xr/ u*f*!3@Zthk(Pq"wJ,R Y;UӨ b'" Vgu!:#42b"0ؔ2Zy1)5( BUDt3 n $wY. ΄Ԩ)vHTDs%wReD֝Uq3CuV[%"7xRab3V  1 TD,)G6 *ʶgx(xXmu싇2 lguu]/=r ~ؑ[i`^q;3C3Һ*=DML5kHVH)|6.w~].ZF f 2aw,f# STƉJqJXb'CD`,TeOL"o!NpuKN: ;m[»W]urc;@͕/Dhsե֏v|Dݠc.omJ7T)LKos0R̖wy6 dC%ݫ4|Wx{&;ԼTrz(٭72ݘwY;~Y[*%gKFbi·?/R~8ߺpϢ[?n^<Û4ӯĵS4u`5&h P76 F4?nBo.Dwxe"( dɢ6*6#|Cl NKXFѥ%noX1Rfƴ&r~Q?Niڧʢ/<~7wJ2f%;5%# -gW[ʗ,Z-:`x`9YO[?J{44lD&I2gˊ+p4&7T)LDi S u9{>N~X;ہ\  sުW͘|<-Wf*ݴGr9k`à7k3i8r8;e^*S5lAŒcuSyy:$q`!7< 9A  \ a@mzԒFsA9/kK`;䩂Pފxp$3jR9H[aO-q(dh[a{>w_Eg>\оuI>7"jBbFLF< 3!.Y'`?8$=$վMtMy`X&fL0 }x'/]G:_WLVY']xfvr*I}QI(e2s.ˠ|UWS^)O;?~i*яx9>;,>F|[ǟGҰ%g\=LRJ+ vCםֲW'SK? kA{JIM.-6ҳU:+zߘt=;mņݍۯs޴ٓB/j7;Dߙ\)P?Л_2^>Fh5ni7^tj4!\f?_{WO d~"swFYLM6.6JMr _yk֗ B#|E+d/gyUCKjcZ"hgTv9G'=G-\VNAU5b>3wgJbކ{ot _9b18$rt=1}?u,ݲSͱSG3\I^3R4Dh&$%EC$K3 H{''$}ӵ&/\ \f;٭.Y#$Et|dlz&J%:j@?nw}c4%c)ɥp&HشV#viM͢n)wp<A$26J BTIkx^"g)cbui"^Un.{6gʚ8_Z1nzal}:$ EI7$A{n*2Վ" Tx㩎F&--Xi#1H:uF&+|H6`rY||E!ujc<*9}k.nӛ<Lc ǫ{«(qWsh!+!kEWԈh|aX)WFe~Z[q3A_D E W-eN.eP-8e5J矯P_OZUoO+ur>=0'HU惽ɧNRjgc))s6_rWןN6~m$vQcLS3 A,b j{,RR gsSQF{Z8A4| .lBAŅlGCF HQNYElsoYb"/T5&^To|y96—SP k6[NUa(-DOŻ?qqD Aj.~,}oz8Y2?GG;it=olni6mL62)sIqjb|'0F_;!IWs15ǮNŶM[ʺ٣ܧMR7nGI{8tw=ћ`' N\ ajcmbxi9* չY֓]׏'שdVd =e[)B\I@3\xod@U:Xv,qe {^KhǹK]ayqi޼]Nاbw!,PK4uNL&QRIO-W@G"C&ۉr@_HelޖbWvKUPϒO&se치Lmȁ+R֛WhXN\!(\er5<s4n27W\qNj2R?s乘LmƁ+N il{sgqs&U}_F翼= gmrOiC-vPfaIysT0䪵<%51) 8YjCPK(Zh!cIX pwo5Eet :~xzZ2TXt:J {*N&Ok]T1W[bn5{9a˰l<+J^v/ J-iae%;OEɎ&TEr{U0\~іjO;j_c{+sL +td@pJpuģ$`OWIolVC}$&z\NG4H6F;9nUZ^[mWELɹǙdB}r 5WQ_7##J^ M,?A0E&&(UgGTWt I#b8(pn>!26EQ@24Ct+^WZ !iz횒$ƹG; Ǎ#*NTD9-p<-67?۟>w ` ~(b1k4!sb۸ȕގᴱ͇zdvePxYV`*n5(w@˷PEi¶ɼ7Q.M(n!P[ͨ5XFԚN^= 29]AK;@M$(͈e%J0*Jh_JcR$cBC99D@r@DE|]EҀTZcjDD+M} `6H^FLjG*TG#u` `A牖4$OG:#g[m)fl^\!Tc>K]z|jO/ԣyTs߻|Է!g-`v8 bz 7ChߘoL7i4}c1Mߘ%o<0`MIL~6AZFKd*e0$ 0Pm?_rݤ4q\`(:%, -ּK\c%Io'=q` S0fWA YF ipQRDH2Ƽɱ䒣i;t91ı>2rϖPoOP4%G.?CMSn}l ˡ2ƣ*Z!]Btzz6{y2#ɮ1?ZB5&|(~˨),(l ݯJƛx.(-Q7ġ\!•siJTS7<}aA^oN-]?wj=^EW@#f( }PLSOKRL*@1G,*|}:-o0"- >J"IOuDFc.AFittRHso\.ҳL1$e͋ U'J3Uh TFD%n'Qԃ X7F8{jeD #)/aIf& mg ddFu"0++*^"}_C Ȥ\ieR㝤9x#^ў"|Āy^$<%"uHm9vڹ4hA91IBRщDD X \Mr-mftARb:bUb/"SjdY\Zߞ?xS`8Ǹ_[֐ dE\e>I}9)~rS,fְ "k-*/hq[{pD5nr`0J!*9N& Mh/gjI䍣X'GM5(s r7\sǬsYjFX_F6PX\*Jw)~+.meD7* 䦖죒g+FeΏ:CԨIRVSW'}70j\c4Ez1_POzM`v T,_χip~zV-77k׫iu=bqqiI֑@m:$47T> U^|4|/ٻ6neUY"MEӢE䐶ndN݃pWJt6983ތ98LWLNQ:*#GOmֹ欟ŨU>ȈsKdƓb1NECz@qs{jS47brֳ_?]2}1/ƭAXlqx|;9>N7fS58 ܓ F1?oJ "̺.UV,X8B bTB%|oI\Kŭ +]o/Nl$!{Z ,uV)(jBG5VфfĔ TP&{#K{W>z@lZnm -D}@]]8N$D$')(_R "0gLu:9ԩbGmM~3h nv;mxe.G;{(-QKw&Ы_2ZC/Q:hBܰ ,*OdWL9%J )4hcyQ}eb+{5]@^2ϒFAaE">$#ur4<S^n36Ѱ@M 1 Dryo%`u. %,wVz=SI(~ƺ{pO섩0|9x:*Pk\Is K*M4!y r*$nhU(Ɇ(*1,Iѱ@CD3.1*/kM; jgx:RԞ=Zdq/v_g|w;lwotr]g[w^ꊯȓӺ_*vٳ5=eO踚v5R]ᕚɌVXMAhmEvQ &(RbQUetHޓ\nTSo S z.II8AQR 52vFzdNW ;mt'Eynˬ\<+N?0+,7 J7?pElbii%ŒRk&uRzhFAwktrÓeƞC32lr^evĄª"jNt9ۏGdb jwۢmڽ23bQrꒉ.%52.רp\RN0I]FR2jTxԈ5XdℊcyHH@:juٴYQ?1 bg%"2-"G6 rϣh1Hu[ q@34bH#t SB"P"5Q NmRrЂ3|0"~ "E:F~߳#.VLlc\=.gp52x!Ym#f##x(ƑHlǾg/xwlc<@هi<ȍM1K\Q4яh扣0G瑉2q"K- %@#.i0Ǿ8p_xOM$L3jIH%!Bڲ:/L%䌔b%AFa]9~ e;7o,Yյ3뮳mvM׳]xo\IZ5bS^:THR/%[Mr5.ڬh7W[rۼU{;(=/\vOW.y(0x:_r뤞FҥYH^|󦛦?-)of5,&{mdђ4⯵^,uTQQ^ ^ hi}T[bzKoj뽳G|1j/hK\}`}1W`I,W1hGLT1S I'8+fS_a֠B~ۭ/> uqO Bl6})/5r>_UYˉ\x %"1*7K`i!{`%0~rQY`h$BՖ ɁAK+P&0)0P tL@8Y89()ʝS)" BwvFz`Ȋ?>1=aW/ӫ p,'v bk친 bRy@#JwBI8`$ W4&#d7Xo$Hr$j '3'cϠGyDꕈ"DiƢ28WFQ%4ͅ}鑧 Kp>3rxgr(=琼L{O/THpE WvPoYgO^;=o۪Ent8dehQR~=h͉H_ijBоvjo1S$Vh#\E(![ykvo<]l6pf!8Z̊pIA(eSAAiW׸uCUC~½ǪuPI]S}ԮJZߤvW'y1O䇜ߥX9WYbF^F &_sՇ˄K1z' ONcO?].'q7rF-7OK{HBgMog8m6O|?"/'| ܠjlͼoh䧺|?-WdT5{?P4Is{B+ݻor z|9Yz1_!.O:߯*y&|;0W8I?O,tʽ&Rr})Uuʬͺ R7N E&tćimf`nJ{0ζ;P`r!& hDHFhAdjYmV-r Ƚr9悦-b8*+⨘bFPX[wEq۸5 w%my7>ƟG-b?:dǝ_ iO6c|0oWb?IǶ\גmټ[670|&Pr+hP/eq<7ϟNe* L:ti%}$^kr59&$w&4hTYtPAFi4茜-*sNC"v*sKTqrN5_)uiW>zTL:W12ۿ;zWe3p"qQJm)% җseخl>*FgUlKPCNUۯqu^{n=uFΆ5|7*_RN2%PZ+KVyxt~J9]r(ddROL9F<. L\a"%.3*0X#sKaxYuGQM7O% "9ٙ aZ4 .J#4XjPUh۩{Acf(d8>f+nuP@XR C RBKfuL^R/1\YYK,_]B_jk(Yõіh ;<8+&PTߨ>|c72Vԭ m>feZI9a15D1`@CrqL0p6rAE]W i{go^Op5-z۬TpƹfTaV8{SՋMO''7jTcqkNE<Us]lbvAbA̹s0JYެ풓G/u1bqqZ6# HmðaifY>Bk|̫whx=_f`brQY=|mAFHX"|7eq*2lZGTU*_qgȎ =7'˻7o?{r 4F;.EWCx -u-qt0q|1nmb˭?ߌArv<>N7&ZăpeA (M*¬Res!İ ZH~P@%T¡'6mK]/VG ҧ=Vs# +JhgJDQ:ZEM:S 'PA纞"Fҗ61}8yƷ"`ZHq@IJ)I4NRP5B*AD& N':{Ol}sLE[o{QKKtgޕƑ#"y#G}yb=^X dXP%bf2>~AF|US/E;X2j}ᲅ͇t&E91=& &;@R4P"3*k2,?DžTZa#c"4Xc2JR%mKJ1yVDLS\x(7T;z9RhIp=/y㌶3PYg` !  7{Ӂ aeYŁP q yJ,Mo'ֳcsKA!q ;\Zq7:5[ Mתջ vhc 2)1QbvƽjׯIPw&ϴfp?,MV=pCreQg؆ksQl(L ?B'OO?:sPP *@NgKcVRN.i3iM-rʖɵBrIS€U'6\rpdLuI5&av=0` VjuIc-W$0erY. gǀ^L@ۂVV`=SԸDž%% UrƟgN *PLB-`m*%PtM':ouΎ[! GBf\6Te#GL^:+MUe PGi'PcfBbLEW83xcQ9Ўn& p*Oh)q֎>|9ndپs}/g_F5w^<'cږƲԾ{?nWM!z]M{쌫5J /F!g*ʕozdtR*jI6.R6='B_j$SP+$ig HFnَJ7,63 Ä«{ؘ8\ne|`HggN/e,ƣEBޠ\Fפ]4ӕB}Nly KX`:k`۩],ɇ޹p1͎}Q;vF8`7FYc9huS@++b$/$s+nzaG+Vt>e A2AZ虐!䀓Ha̙1vB2.Ygn6bXZk,Q 4YoN7[jdmD@r:@9㟊S0hk:Y}7X~ަklvWу842mPEz{v}ى/fFv!qg/!K"&DX0)X lZm%x-n0~$,&j}xe2|ֆ3CX.7$n9w -xHU`S$WI{\2o&FҗW]|KoБ6k[-t4F&V-.rJ128`gε.NJXa]x5:sN{(~9n="&tPDɊC;TV/>p~.f`o9_~p,:.ƾ ۿwG^xX38J-)򹂍*k0}hI]Oj]d/«_ZW'MXQG/f_4`Bo?leh}BCL! qRSCj3ݳ߹^.;Il <gY[ՙu5Xc---/&qfgb(ҎkJɆuV@%_ ؃e8P@2m^)>&zU.Z -y?j}q}3/ں8NѭOO+64qY?X&r4%G@UКrXVk\s) g"B`Rd%pW0q^) ;"̈́n9ܼ'):ɩbU`0F8N֋WC3[!bJ8Ej `UDC0 9H89s=j;򔾻@b.yNITp#.qԻѺ[{Bt~'iЉO.oRڵ1陸ۿ-y9ݻ^ku_y}REvI,&I\w3 M޵ޯ&ԍŋBn,ƥtTOD ПR' t~9z|79 5UVm苘D(+kbXa*-8[+ u~nkݼ94I٢&EG?g)޵q,Bi//7>$ٳ906K5MjIʶP.ÛZ֌5Yf׷aaU`Vz3/jN= _ӛWބhz!|+ _a807tJBh*tHW N* ]cU3}6~C5U Q+\uC}pDęgѯ|h>[lMTcQQVQ1"pc\(Jz=-N 4}`^Aha/CӇd[4-i1]O*'BCo*}tQj:++U,CW ]!ZEh*zWHW Bo|>cRv%f9㣑a+; yۺϱu"ŵ"``7?3^Οȟ#AH62\h:]e t JNI62Q2\BWu(`FBOVd}է7t h:]e1*J կ XUe_ *"NW%HWZFI*VjGUF;TQn\UAiu͈3X)PTo]#&G$ȟ$ս!iDk:Igl WIFxR2}f_nn!{|CjNV>&5EGilau_&q_ZYHy~ !Rڃs\ʾ*ڼxe" _K[2G@P/s#n0ߚh=\o7: N,ן~x{"R-,"T'r™*1TOe C/#€;yhOdE:y;rrӳIP&s*FVƆT6sDJ[4V-.!hy)0cQ-,c*p: V++( ]e]Rꁮ^!]19=+,vhv(0]B9b=ngu&yXI}vs?̑ qggm]ޯ1(?'>xZFO qQV:dqCMۤo) fzyā0+3oF?_7)j=d  ց޲:|x$F%\Ff6@rkXAĠ%N&5[&}"M#Ȩ-ݾ~T'ȣgw'8&clN~K-"fxEMQoQUS٧x85.BH)AG)pRP% JɫTJ gzj}ַ!Y$ 9*L,I>z\bf{$%jQq)FdL;^P$ G.œF8!HE0`ԩ=0"ALjd,2#@'DL(NR…u(u)S:.Sz~Ln{)o\Z_6~TjyR|D,4zʝ□Ba95/)Qʩ'hjJ 1*@р5A0FxFc1 `S~bQ3gb>Q4AZ^ 8' 4QB҇*&$@3CT+3),yrT@ 9YֳP y\p]b!y\>r~Z&/vO|1ۘ^sabloGO[*5R'v!.qA!Ж"!pIc }44;_Oť;>`IcQmϦ<~Ι"VޠaUvW)S]M3c0 ԁ.BcmnAiwKbw<_Mr>(xMgʍUϷ4PSj! ˄&o#1/ۋ0ыZD;'zˋvnZcĆA0mc5[[pi|f³*4-4> XhiHHr; !$iS!J(jږPc E'ras)tT\:@^zwSvJ>98 yժ_nN#{<ޓ9rHj✈$*z$A p^'(:֨^>W|"ۍ-66ƒwc7o=mySya\uya׼nYMduCy½.'7xoR ~^+Eᣟ8vx;yly g沩ImlaTWt,+t9^^޼0˾E!rӋ2uJ?zf?e-%,h9YlJp7l͏Fy ylyOv25.GIش[9?_:c o 7CF|hT')'p;{_qsV:B[\.yV@l 3Rg-iS<tE&TntOZ#P Y]٦굩]wϧzL'=}sPW]Mfy q:Zb@~!Z]:~fAW}tyڢe=沾u|?[O[ݾ;ɼrw݌ emo-# ma/B HqMYlXvRV,4Y@b $EHeJ{^ZF-"nGniN;)QXG2΀̱p1|"N-&(9DHt-P1HL4rJ (w[ %QMnQ1Yxk=l.oQL8Knj BzjL4KA" 8o"G3ASnR($Cщ(ÈJ(*1, e'b\9T:R'j18'j=9q=Wm뮜'ǞG^Br~}9QIx}}pL.k]Q;]jIYb=[N>K%L+0` \ LdT C] 2$A%QjBHΑJNj 'z 8#3$R*&$X nf,b#X.ԅP\>7ަ˫(~{kv_N퍿*Pdy2^,qƶ8-M`Jp4P b2THaBI߽&$(KPoqeNyllY$,&Kp2 *mGI L#9)]n6 &fWX8Ma6k1$St9N@ G *)@ U(0' wZP@ 5C hE{rMh,PqB9 ?.$G:*k a186ˆǑظWccݵF|-h=18y-2h<!$p8 (tpEhJTqIRQy5D%Uy_ͣ*@!` >g/̈f!/.7Ylˋ0/ҁ^6cBs;-!Y0m#f"#8.i #Z{/6>,6:C^ J}F 1x(j=~C~fGmK(樒w9S8$n*4Eł'I+MP';Nu(,B&5'Œ!/a2&E…3&(TuJXL\>Xk-32Ct /~ r+UHߌY^" SZǔʥU rښbYlMۚ%%o>y맻#5~1cȝ!8q6a7j ]Cs]uK{zU)!`LNYIRN4"4׉GRyL@'+d>VC= ~IN̤#`A bppȟ\-%:&}pGڈ꜑=4`',fWsߦQ率?-zp~t#zϙkXOOpN|7 51DE+Ye]t ̈́Rƈ: p+v(fPtA!n18{nL=TA4Sk)Ia{IoAU" ɝ QN{ 9Gk"'Dw]Ee_1BjMm*lZdeXT')U1Q)uhUYoX 6@Nksu zըKs,rM\8 q IiPs)'&ꋡQG%+UZڴe3h|@ePǺl]|c{FҙɯVi0_8d=.~Jcl~$³ٻFn$eqG2r&abO1ybaɶlK݃dnE6YOG 8qן'-J;-}^la E+%kkلz.+,v~v컃Nw6B˟"Kfom#߻NCgnvCg:ȋa~N7.hBynm]`0+O~6x0y6&rE!h8n`L"pހ^s x6jlth0CݿQ>_РcKA<.X P5oy}<| ~Y6u]4encqFg+F9Y))B%oHl$(J-7SGoYiBL 'H 'a%<q3hZH|R:_qz1.3%g|2tR|}:G j' uBY)hਔz~ RHIQcd"c5C&hTPmlGu;%1 o-ka^<|iL3V5yN#T@! `k^DOpf*-3G+o';Co/ġuK.?⬯ଠ/|7oPZ!N LFg]l9EChʪ-R Iy s)SW:~ ;H_/VxkKNx#]s=uu5Kid;Tc#{x4&}<삳G$)`U( ,>5= g2aNre6l;SԤ"=dt1C`|:K! s:20 ,^QV.H}Ib(]&SdP3!T"'sN.QSK{o,099C,߾9B^7B15|~&!8X`y.ɦx"k캊J;Q 4k*PS}A]@LS|6$z%YOFgCSUφlJݮT)zWH=M>+,~0@qCР@ ѸZyK܎p9;UdN9G0D+]PM2u:bۙ=c"K@0)#E~oe%vD2i, K!鬅fd&EW/D)zEVCX"rPjAFEaQ"CɉJtaٌ.dYIV%];?]:Wg>8%OYlzlq`>'x }yuh6s0㯿a7!ff=LA8?:N] -#Yu cG %w&Z|'lO15{Ϫ)Hg~mdM׋K4%nM_y܎:pRŗҼ]_O}iy> oHD~zpp4HJk{NOq55#)mF9yȣw0M};Xz"¿[u›e`vEœ/Gr|t~:.vih0<v>is[ qBkږod]3jm3FofyJOiT/tz0?|śˎۃxtO\w_x}߼y$JQNKnj添7yFeg~{<tyMWǑ^:Gq o_~ӚMoNO*J{P]/Via'׼&#,6\h1䰴"""Ƨ7 dcr R1T %H=/ ṿzX.[]by.f3n9{ךccT6I`/=Q$ r0>=s kz7@4%C/,455EbL]^Uډ|I;sQcLm=l4I'O2@ Rf.E"5>kg __utfm}W3ֈY7+w6lwX{%PNgT\^"*.T\5rדVIů0ؒ09 ٘*b报tzs U YzoXY$0WU\ \Uiv\U)-+%¦%vy_y;X=nMf5}0x"x?t2}4hgFF]@9 T"O?>'34KY*LWiu3Rѣʗc_H0k]d6/_s#-/L!&i7]wVP[ڢڲ;WRPJK UJE,P*P@ i(lr)"f})w[=3BɃQXmkX*d%m)ɘ()gc/J%>'%.Zҕ8GNמf'BP"j} e@$Oz_aЎ0f J ]1"]UKb_l'#i!X`[rE:A"Ϣv BD鰂MѺ+>lɠS+|ISzR9IZVk sR>XzԱ0& JlӅ]lbcd1 _Hj[):C:6SvmA^q;^o!|7{c;P킧OT|AZ t P5v| 8VKgi<&Kn*gn:gn,g hʎqXqFg+F9Y)*)xȳ+e6je@B[Nw k礌Br$( Y-()dK !ȉ0H>.t+Й8#+Ͽ*!Ŕl g.O)'8$'eK .=::Beӡ8W8aeoO0GGm@2DL҈c$X32A * O#vJ%N`sI&%"IZd/AH9Ecׄs;8b8:ݱrEO>m~aa=ՇO6^V}]aw5yN#T@! `k^DOpf*-3* f֛iؿ~/?roo 6ଠ/|7[o]NڿIGpC.»kώk LFg]l9EChʪ-R IyT= B -ϓ6᯾w֔}o:pDN  xw1JϹw"'?mfءFQfb]AdLeޓ^g6M9gIG^;=*U-Cd75Id Jcxpƃ7Jdt3q5˻tT@;*Xy?J8-ݾ9B<y:K#61o2}iW^1{]mo9+&ŷ!Y`簘a_m!$[-$۴-;ؖEVSUdS/^TWO\8;û [~tlO^:951J`KFy4f,{nƪ}3cŝ,O\x!gk8wHoY9{z$7W5c 2QQfLYpcrYDm`8 &2Esfl5qv̒{>징t> qۇ+ZkG߇ظ+,i,H h(6@04*_LSW,EF(L 7Y,ZG"Б'3fO`6LNk %gSK|"2%-_\{cj Sb,_ȳF=A8]? Ge߾ #UFq@ucFl3X˒Ճ8xޣ$SQo˜*d4/Bдʸh|lPԗ7>S1H.URbDdb  +D72M0z!˵k{֗-*—QԅFMѿW.JvS␆+?wKbqǾʺw%5tnϿioKuY f1@Gl<]LF7ɲ(cx 9ecAbpppt'%AئmKiV`=w]/KHSZssqqz~.3(ME|էGF[Kw%N] "]wo5uv\ѸGSGh8#?_Qѵgn/[ݪmuqƇ-aqx!̉>x5^Mt6lw1432ד=n:Y>4D#XF|=Ћ>'冇jQ+ͽg%gR #ut,i<"fyT`lun5F͉Y56.>/T~~|_/3?ׇyO+Οh)>}=F3_wZ`u]K]ޯ<cufbmHO_Ͼƿ"(f'^=)N*' ZGq䢰thTX8-Tb8oPض߷'S9atTdvC_+vhzZv;fN5R*mAʠG3rQ M'w֘״>8dgEt.qJT!9,ee7.& SX{i7 [nt8=%4ޕ⾟o' {L1} +u-CZĵZh]IZֵH bui]@x2Ov1uQ}VHJ/ꨂ&sB#o``&NfpFeމzk|T}På%kM BQm^FDt'mĘHs-J k9>Ywƽ5פkySA1 z,כkH7lWk yBX^ԛ}6W,u;'_v|hTҍW(jLT Ge{`a2J12) %v&H@N$@uϐ,Y+R |i(Ҟ~L~syq$K+|M\Vvyrd rd _FKT3a.UD` B4aqȑ4ȟBQX- #XAK[,FI0QUu@J?,'$)%y="rjqTg G>7$@95Z_dɇݨHfWv)cs3E%M~yצ1-84J J'hF IpS6(CfoY@t,;h<8 =9W'HI24@R7)"$dD6o䩂<\waRэ#/^ǒj"f4hYƣ XB 9.LMg F-^9{HGJ*uӢGgw?j&Խڞon,!+K5@(pw&B(Հo1DJrbVdcB%(D/\ʙ]8֑(S4Wg zj3z<4p!ZL[i1Mۀ~:* G:2쫛f.{0wK ՇܥԎCx,;:t4|2ÔĄCoj&v펢v8h{wO]Ak5 ۧE}#͸ᅝ?fFj9}`bWt_[Z]XmT$rvzc{垆 nj **\b 6޼e:EiΦ_eڃD/AOrbKy']O@ݍO,R?IizkHwSޮŲl/V,&9q:Z]m;u xEnmn\n^xv>-*@")\h-teoY*D|%3Idi; 9Ȱ=Qe{b_4^8M3 ?7Kf$x#Lɗ `aQ1S 49Hz}:jU2C>j-x[Ldze>]X+pJz!҈Z!1&єxy-)=ͫlG 5  P!oYXsScD=)S*sĽsCM%gFE QJmbbL7@S׸ O&#o^%pc,Ѻ>aل.~5.Vy7i+V=(T: )6k,z՘`b<(|7sKwZG-_JL&Y5&?By.J&$K,jRSIofQ$p2 JYV&dBi@!LF d_19<$$$w2zͲ]VjlWF"6s2KӼD9FrPm<<@W>x`ٔeOyUկos,e3&2-g9y}T˨CBHѢ<KB+y4Ab! d >2s5|creL̂AO IPLt+82Tmd&vdgTj3Xh*cc/Mg29P^]l^/r Lf&##`K5HS\RǓ{xN^rcP:ˆ4 s{P`S(zUFɷcf@ЪYN.lVKgxԮ6;ڶ2j vcF LAl+\&#g,2!d"JtL#JK.x#9L!GH&&r!Nd8F,# T'\e<&vƾ̨I[;as#n] R/ǣI/{3?_RFg Vv>ZȺ [ņ7% x#z}.>i FS x& w8#n>W;Hc3aP?\㏣e?~z2u[w':!%rfi:jʷ~T!Ŕ=W):X癍&37ъRp`V\唅{vqtY\гb%<R+EHL4úZa.A~Z>[h!R"PYmeпЩE]Z >(Qg~88ge 2p3*1I*%UN'E9tCyYo|Ŵ^@M҉o{HO;߃V^⛛/:NRO_u^$NV%\\VȽ4;@O*L(Hɩwۊ΢%Avoͦ \)[V+v'>O<!m jRiPMƗ܂CTB $KPEkWx9^vK8LYrd[H ngb)}n;_clI\&%?%7/戶+K9-*hG ,=l³~=_7[IyFK cUN6+p&ncմWZe9yS9qeЭ[((@)sh,Dv%rƝ<&N$M~b -Nnj*-:yR%B*ȬSaҷ/nG3|HG\\=S,_ޜJxW.g9yՀFL^?{ n^ ZJ_ǵњm%TJWApQ Ԏl0ZϤ'?{_*cE[+XbMn\ y- W.S󃬲p-Hy@KE- Qkmrb,VG-L|*B~28R(FA"/sch2TfH58P~$*`@oO;<6 U,~4:T 뒰LLI {3A5( xPIZYBt,__cO5Anj64O,$Q2I7neȢK.3 N?b!`1*.ӂ!L$g3CfTĤI\q0Cj~G6dUSj^/}ĿI_?xSIp`Ǵǿŷ 5oT2λq't ss?L&?ao*w AIk4Uf >n9%UCW8b3z!-%5^Q=!xqWͮ]>2citL$Ius r5[?6i 6돋qD?u7Jymo͞h죊RdPŝ!Ͽ35hj9}Oǫi5[b_P/+ \Mg[6+1KJwKAp4jҤ#[,ou#Yv$Ptnv,HhG~,Xvpp=_fEar y$׍v t]N3p.#e`E8G_^QF yFmz&^_;w}]opJ{zw7V`dy#Ie,g vZB[Cx󡥅VY|nIxZqur1nbmHʏ|3LXʞ(O9p5g`XhOWliS#~*K ,^FX =>h2yw$nw$k/VΛ7GN󃭯{NWd4wZCd]CimR Bā' 7vNSG:j!Ff 3lu%pVBtYs% $JVݡuj\]JxA}&V;3xރ*zsKK&u֤K=vU!+2 \t da; YقE gAd&sP48)kqUMy_sx:ؘؙ=oL죙>ԣ΁fGbU GP:KY ֐Vy^!\Ad ** U:`v@YX/w)%YR):G\$9sY& qsȜ:=hr+YW+yf1CQA XCf^\? ٲ-wzIW]xN:Cڼ v|։ih`qAfUJVU =yzhќDe"LƟ zyʪ= L D֞ >^Eis@ANLZP0KnCBOq\8-<V"Ȭc8X]G__CQh>qwd+mDVh`LOYB~ʨE{:dy(f&1d*B:VuJf`Rp޲R!lN\;IHzQ،#ZzoY[#gkgf4Ku[\{u~br_yYr! xB~90 ^ j։CW%*u0ì@@XSLՒ?()}3B@^J . .+ɳ9t.4\iRykh;'𖉭oPW#~&Xhz7hU eVI)C(D[tJ[eŷ^|>]ںuIu&խa-s16_XͶ*pOv5ӿ*Tq؟%6Y`e 4TarteMH={=!sP$8muAdC,&@`d}([^-y@ c=VYXf%)!EEf)9S*k9x'=Y[FΎlUf{X8x9}yT:ґ;V K$ڰ$\\̢* ,Ԫ"0D'$ *#_ <"#"w*cm5r *#U q\&/?m 5>yf={. '[o ([os)=9Ke9IcTYs跩whaXTp O`/Wj/lBLbTi&ƐN9v?jkrJ9V긵)K.JX0]c974sgRRmkٮ۳UZӅqCumYN>.\PT6*$M/6QK:*6_  FghۓX$LPL=4A~HAyFvKhVfe(Ξ' CR).%ɷce,'cN*i;i<wsDL{]#MU,gL H6\&"=RP"" !ض5bkl׈ϑ^u XW3:Xgk\P[֋Ӌ^6e@Pg.o$ X'P]HYW̵L/>l;eP>L> [xEG(_2CF1E\؞e?>SBG~@ i& Q@%Sd*rU%Rd3]cWܕ1h33d Fpb6";0Eehzg9{TH 2Bvk(`nӚ3:f<ٻ6% e#}0XǛl6xI>mO+THJ}3ġFP$16{uUu)ܼd9IM(ʵ& BNި9d +:S|zs03lV2=y6e-+^n7M獟̻Ԡ煖CJW_t17Cسˬq<{N0 %vg[~մi)B|k&.ATBst ~d~kq==iL s+͙%86\yjUPz<:M]p4b_S-^"DB- r "d ~&YmX$9&jš(*eƚP6moozE8G=7 U86=;~^ߦT dh \Fa"~d~!rwD (U"9J}&JTk Q$ש$fL XByd@>%ch 8$)S"!wK9 Ha9zT w.MyiI ɻUꮦdh=d,_aP *V`MQs%5Lpe(^q#.pqGAVVGc̠hP3O;3M}=!W<l M@ 2D*(11<0EU-2>(MQ'>&Wab B]R[t}Ƒ~uq#{inGe=TFT5(A<լ3hGӔ6)7 cl.^w5a,(щ6Ud[H9sQ{u muWSWFAhcgJh(Z,\D,ǙUň}`P ©DpghH=>.omS7NKWR-\9*4 0M}CQNz€8Rez^ځ.^0=cmvvڷ(=Vp񊢔⏟FvQB֘OmF!XË9B\AP29&7ܤ!R]|1G}GeDqKf[FEbShɉ<K#4K)cSM !$ĊZԖi /_ء NB#3x6RYU4c,dP!#Qp0I~7tpAV1I[ ԙcHejQ*X}sRIlHၱkT9V&%-L )&]5mRme|\;t<29,*AAEATDQG3U \ V[cR;ĬW) a 9HyKɍUe@k4HH)67V _95Tqʻ&Ve_ړ[­l$={8vсTr^Nt&\J"$ DJ6*Y&e] ΊߟMo>Sg)df]ǝo~JFwDQ~FXH3&eQ|VwK*˩luA4L9g8l,շzNlZUL((D.2++v~Y"uT U}Mw&wx ÏgS;mQ2hjdע8`7fûьpyK DN`U"SD-WJ[+$]%<r%WZʎ 2`Wo4!D0'W\N j֑UR"\1 n{4%6Y[9/͵ه[]a7.mq$',CKDG'p2ma68K s)"2ւ9cMzb 93:xMx7Ab3 ~|F O/"C_F^F:еދd tz&:MnۥDjO0QDVW@-&*Q)iWo+ 5 mU"SD-WJ;zpE)*,U"Wt tpĝ2&J`'UxUT<(t~٭Ӷ R0yۡ_%l4Y'8S"KGsf ε "W^D&S8RO \sx\թ :XfWv}b4w~w,Z<4O,[R{kY)@2{pP9 ~~$L. u8ӵp_gph\6wgLsmٷfb3N'yLEqoeq^(3W6v >؎'9P-㷽)Kv<[F-r=&ه_'~On'Khm0 qVF.WabUv[@c5o~;ۆ/F,ӈQ܅w]{ޅw]{~Zͥ#: rw»\].Wx1 rw»<].Wx+Z+҈tiQ.>G_-C b+$T΂+ +=]iI奂0NKO99-UkPC5nlʦ p.}N<((]ԖHU Uo<=^kL< X RRd\"sP./BxфIM݇Mw$}t|QKذU+B3!18CFXRPi^^hX)Q1waʽ3ø)a嫂i9c*]L>V"xf(?t7\HNO>cKߛtZg53cN^=4Qr?h)ٯ-cF J3QyThlF2(Sy8:.n9ִW:/PBfHxyʖX,wp(HFR#8hca5hAh*XʜBSrʑQ‘݃怺#)1,( JeTAXG$8W 8?%,.d#Nx D7"3L&92D˱Y^ AzlD8EI% Q(+a߰6FSx-gF#Huw6$a))҂PRV$e 1xF<+|ŕ! I-Ku/vH&B5Y5^̾ 韙)y 1R0F/`/[YWL*ENva"2凟k37!+Y`9U,ي"l_αE ߢ+)(ԝ8|y N[lB kta=jUazS%SAa0E_ 7ݚDŲ~[̺wy9r(T,.&k ^Ydhr6Qٻ6eWrKf H D0zxL z8\ IK-ǀe5=_Wuׂ .qnFZ{j[#~l~_|v~q]cKfN8Ocjl6YqQ tEgO n鴫ٍ$ݬ [!' >Q̫d<zt8psr|cluq]v+C9]ӌ#cdڤ>`6gt0w>5Jl"*Q&~~xvOwo^&}~^ywG޽%8EݦXP,r4\~ڽk]s ;t-Tɻn|{fm~O=hmK*,fq~%pU ? hvb%tTS ߵ'd2ܔ죻8#|F$(k5RB]2& &A-I 4p@ bo$=nf; 薇:Oy!`<"SJD $H +TZPȡNggwN/h(&6Ȏ ;ՃE-S?hiZ2e_z%A90ywgWWvFcqR)(aY1yY*𯓞80]C2G5W)Lܼ#><ޛA=E kU5qQia|_@)VBS9L`h1S7C[:6f3+T>l2=FLqND$4BhSS:1C=eפK.OrR6}I*zBV^{$U " J8[ W0sitW++(k(W61jo"x?0I@"UBWRhIz?q8rr+fXC y;|Jh ':+HF1%qoTsQi#xidf/:tSLNnsta<~nru$Ch6ңLr]vd{m |1}3L~ITޡ4"^0Q4Rrz{Р=Z)HQ DiRBr8nuH\*$X1$R*&XXNW)$X c^pQy֌b=ki ̓?oPhq4|B L .fJ1RL "Q$Mg jn[\Yk!{;8˰ɥ`aJhm#jNJv*)vvqX31ImQFmӣvn n&YڤMMPQ#8J w1PsEK(]zQ3VQ!ք" 'Tc.A IѨʖ(v8!XU)bŤx !-L\iFT(|9q0owINgPQ%xdP%ND pQ&eR؇1a{L70Ũ!>) 1x kj@ %)Y- 1AeSz7nO!ê ͬ*;r:ɍ!6ڪ3ͤ3~l]a[WغܺLs\o!l lb [.v'ixQ^5N嵖!ꦷ9?wC.MϖmU{NvJ.Esq,.yHw&>y do,6$8a#ȃɷCɷܢKϷۗb|<Ǐ,)肕(hUP~+pVEY (!}ucQƢC2 ׆!A`\).DeT2R)|XJP<nWj͢vTc4tmyE͛ gM\C_Wu,76=$]%̯ܦ"'!%5SOXM $F~ TXb`/r?ez`x`)hK?Bڨs3I1T(xJhr 0QB҇*&f")NNΤ 督Q1 tbl[st{'^Ԩ&3J;nJ_tZ6OxxNhRr RhqZpqIc {$qm:Ѝ{4N.l}|CTdϬWĝ! wi6/wdvH/1o&]&:IHv=.fGV[h?_ ͢0NvAqxfQ0d$sxN$\(wa-] {Edl28g &ggw@^ jX[8ƴ;B͛x|7 ue(vTl8z1x}>ZsBGqp49>WDүE7΋n]Գ 'cr](Wkn.yjK֕ۄ涫a6fR}NE19|X7Ǥ>Z s7y(>mMwB.wSȄ;}^pD.TFvAm]mE_c -636/,>J@r|".gKkV,x|m1Xa&"?~GuW`# V_ wݿ~~r hW_<&~2Z^4^/No,_v9Qd8,-Oo= 3٬~6_F?7g?j n׵˔׎mًܧ=Ypywh_7EO7 L==azl rܺpL=BΦËVymx,QP)6}\Kb&s{s1i#~1}Ae~ڎqfyp#vv#7/<ckx@KDD3Qz,a:H NhR3/r"[mӎi!>d_wȀf3e_SMr$\E WeSՃ^} \L6n[SGsK^Xu9J97 }@ &CPfWL}K3e.M!k*2F \!C-WK!z R}]YiQ|o?ȇLJDĬildDZRv/?o<#R%(sI#ڎ|Zqy>3|!ޕ5鿂ױdYG2Zzϋ5v:I@Y&AqPjŌDv]Y]U_fe}~ꍆiSn':E SVwLs҄^I{[{<]FlWy'V'B]חXKprbbܢ; A`2OWI[@tH>*J*0Uκ\Áf13̐GCjECx,tfvEʝ—ڐF TJzP 7Nsc WDR9yY&LV ,ctY`Fu_Ai91%kY3KVo9Rޚ8OGy}7c1`BKiRY.%,J04j VNTIGf449"mWlM ݾ d)GɀҒu0ƍJKo"hSf $͡29Ly3OX Mj|ECBB"_wVɜ2n 䅷” Q)2+d(RtYEnȬՓt\RخjŨ&fw1XMpm@')oe&D%M!CŒ˸#a|2 f6D<71 o9xۚ80aߛR>rvQw2 ?sk^7jW'Ή=kfCGkr׏pZm|e&pҜ C?%TG8:,8ʹJ ,>Q,R)B@ 3$ 2A yVKҜ8.(j"u99ߑ?4@1D|H֌ND^(f ={F!+I돷*$ILg:vw޾fޛih8"!Г]rs \٤+@_h8;R\kJPϤ#?߱NyڼZGf퉣QnxjzLW)C\1ȁ<*QTtebS-iR%+a~<0?_xԈ,WRlP!^tA}llk!Fqe{Z0; ƾui yڙRomҫQy[Z$5 H}tdxYXc[i4-_R.殆O('?]aoJ+G)sS7v^r=wg6 ll ? G fEJӳ s}rd1QΜӍ'?E!~LL:^*Jl5,kG.f7Tx¸=zƣI?e=*Z@M,+ U*@$,RNi]Jj#H)Q#<`-Qld,F9.88R`X"TZݜFnmY En̽*:ڑJ-oN&iwy6eSZ:ǏhxNbU{ZcϷ]1^} *+DZ\Y-m* K4XGV٠|=pfҁX쑳XtM׮}FY[A1D.>a- 85FKU,Qsc(jh~0,)>d`S<<ݤQp&/!hHpZJŜAlut7'6,u $e`0ɩij@X-Z_كpϵ!4w+&58Ⱥ74wtOk0JI+Uެe!:?w|^%tXXm#ydYE$Wl߸> *"MYgHKj`OLU\Ra,%I6Č@!LhY9`fh}Pjb{wCj]ܸt[. /M%c`Pg@?.59m *˂O4so{Wni$?5\ 1<֪ =9{d}M pK@ L1 8a'FKe8qI/')M'i_Yr@ZzcɻD daqgdD ~ľ(uTK5aՃggwRt09 }\%W>?ݛ}:TٝonW aq}y5{J7 ~˗;H?cM-YRI#}fDc357,ohc`ye _/zrf?}sV{E'Zm+9Cw6=)ia#aE8GNRo{e8'7mw675걇j҉w۽п"u#MG?o\~Oo/h8?uI`6vMKhijoѴAʴ ]\1cQ.ۭ~X.I7dF0(1Z UsVɽ Mf~ GRT+XbriT]~~SLyeom_E+vHD8 ]I:enQLnN,@?F{.s`QˬT\msA ]*{[uT,pђӐ K myڗ? ӹA4vюW̛',4ZO}_Ӻ6k"~x<yj5Π- -f}٥j+<@e hHV9^%c>& }l3zygZk`qdɢMA2%7BGY#OcȦs q$52ǎfKFh 4Ik +!yI4YWkgVo6X;w g*:Իv,uέyw=t=VCs 2bT|500.,;nnG7NkS%D) ;r&𧼶N;*sDj/D֎y/rOׅSAj0PS( -T%>&GxӠ3t`ֹ̰``kl4.T7/ 4/Of}.ķݮ;tɷņCoh\^>RaQ7+r>ce)2`NHTԾ,řTWd -[p(uB5)D UpaqcPAKGFs-[pkJދ񪃟{w\ࣃE Zq`AUTz@@u9BcWqg P!K\ŜqTO44X+Cs6M&)v.HZΆ" +]kCgRq}j]i^Ju`0JJ+p+t1W!BW"-3[WWK?4Azq0zmB=L3i)wBc!n?v@OgT.`Ta؟٘Ty<,fмb A;;XL {ܙB] [8p4'"(%Ȇ&I(LFH|$:UB0sڃl7פR* ˬ4K9@"Ґ]#L0O)R/{Xy,tmRU tQ&K%)ynaQ&>E˥,hjfPJ\d Ie1"+^~ɉCOFbT^Eo Oۺ&ӆ.ǵƕ'\΃l{5E۳nEKr@MS{%trZs9ēJhPWC?欼ABÂ3@#P  ZD EL#1Iٻ6$ i`.68_bl6_ncICYW=3|I)OXLUJF(JH1 l"=ěU`̙R191 qƮX2cpX8(2i/Wܬ7׏+Aiଚԁ&B$X#%չR8/A2Jz7nO!TvêuIwm ]]D%P./LRwz=W w1xu O˦a8ՆNgm3~4e -{wnxo;ig[..B2ul05U6]6pA iB7O] -TU}P4w՛L'jzhH7jPQ>."%-}AyQ7ףԋ3x?yך[reIE8S" FKf . T^D&l׍EypهnS;о64i ozmcD ,(j}F92߼I%6H,bFM5#nV/JO`Gu*3۱6/&v7.qKzoM%}dS PR E.aF{(X$7AK+=/lC &ze`G $s8LERB)IfLd,sg2D 2E4pNJ7bi1g0l\"> C|7/OҔ~@w[\>VVc|MQ:]i楆;g)D͕0{)) Q1(*h5r Ry{< ==<9a>`+X0 %plJ7M$$Ŕ!PAn& Xa))j' $uݸA7[-xqROgQ->Z#y,.jE :DJn[Mx4\^I|ca]V-:dv[#␔MDI)mٝN='2ēnw7Z7٧q&Xp\#ΨJ\T^∥S$zo53 GBLw񝲡08P 1 -cP *&+05+`Kc(W}}I>cݓ.v]vAWzKR[]u-W6 ;U~풢 vq\QlÒ{V@qv蝌..j'ځёam׺(S7\l$T k [eȦ^m89?^~uós(^>%ǎoA9[]([W*}[)"oj B"OT,i IzfsX h`uyM7}Bպ4໿>5,1^)Φs\~C_Vh>q1Пu-ŧ&E&M|/PMȅ53wMeYr 6C~j0@&Zm)fm9_ ($rC\K`x&-]W}OX3ֶEc)`UV6F 7R+M0A3eZy|?%-~E:Կ67>L͇~hh_ߍG fo4<ӏ'sj}۴]7y1|y~63J~Wrt5;Σ0/dao\co>]LPCZծkŴvL/8=Yo_8gۿ"}y S=@5S Lug=X A˭3tM p1>ܪ7nۆ{&:Jw V[:um[GMհIџhziw杼 ;Ɖ[GMFX___xAZW 13%4CBFvcQ^ *"A(%a#VzkCv=;îtZ+<UtrQl*-{z]XVaryq(Ʈ5uD:;ƒǪ# tPJn`3=R ɰr[./fJ`̕g! 1#bXt% ҔkR&qCcOi82<`5>REFp1Lq[Ǝ=n.wP 0ATӖ;)w1_^:Ӽ;~m-Pa9GYy>Gfz7#`)53.'gdQT',*3ve6cSˡQ=L1 Wh2\#Jx G6K&mCOx\܂~7 osʣD.WZ!p=•߂:m_htQ7ד9sV;0llH>L 1֨ 11B,b  MRQ#rXa<ϰ*B|X.lqai5LkH3mYwhgQ2Yzl$]?wK'rc&jn-Oinw[ ⡿l0. g]f>GUSGҝ مs @b3w4r-# dRKJKAfx޹s,|_w>jx) }]j./r p[c@W!fJ 68R$#Y@S_kn1h,Xe|BOK*g)i/# 6.v8}quڥm~C( \S7s|8v8%Wr೻C;QbP DuXFC q*UxϪ.)׫a*KEqeR*KP`_ D\hB赦=y/֤I Nh_|Ćb\1ǽ i0šRRa"mPFHhX)Q1waʽ3ø)a1sD6rI ^!eԻL>]LUY[>$̧,vw׾rfYiE%WI_Mbo[Kr6 *UΖ U`hi)1H4{tg;i{zݻ7˯ȥT0i=);Q *@#4ԠӁ4>5K,+`8ykgFOY ƟZW!S##\[`j q)?VqDQLxH;S]0ڷa: <\Xw 7/iK;+޾y$qY6ɳI➪40.`O3X5^r`)RdKJ_+Rb[5jCY5`V0޲T W(l"X+`!κSoL)-4:bA@*F`GaB)7<(S3,"@FJ) )$1b͊Mz%*;_i틷 >ChN MNEW}ID7j4M]饲q/EǸ.?s16c`\bҥVpb#Vl{g'cgW:O\(w> O=cF@1gz4‚S.t'ٝq{Mz3iOo"uKm-V^{,U,EvDaӉ`M&lKtCNT.'}w6ͤT(C$2"*d .If"7I;C΋CZX#S#ت ep#6Dail:1Q]-W##C]ݝ;RԵ-fk|i.'MRQ?TY-0`6˯iPű# MKq&rbДJ0VYE @eF͝QFG?#*$;A UtdllzpFQp~h[PJl[cWaScv&k4A*Gz 3CTI80e%QG5 Mzz<@Qkg`#Lz-YH'JVFC@ޕq$2o dy&Y a#}Z)RP } )CRTK1lfwOOUu0҅ r|]@/DLךUOhog'"V1#br$41XGn~,㷇 Qy"e%RjT*m([\D1|>b;E1|hA%z"׏ae[=nRh@{/?% 9VrŒP; ""u+MS&aJ;f2ʂk [R5,m:[|oD9$"/y0\ [k-,̈}vrG۝Ngtis]r3al.9G@DN`)qkE)3U9!"EʑQ‘M&@$h~t @ XFǤ"9Pf ( РaXGz^LXTʑ&qYNz>0k#Nx DT3&k:2D˱XY^ӿ ,䟪 T[V)*H*vVd4PDA^Q0$2k)u B\=! LqL"LkQ@[g#*B)2! 0:ˍ88{M%U*ܑV ?YKgnoJ7d<CXÏp/L*uAvnz9*~z83OC(Z{Vj5Y6TE}cd.; dB#x RL݁pn u^^@0! `-85O_úim?M&8۩E-45uqaj[ԏS_BGG\W%fP* ?m sk4H]߮`8#\.Ĩ6<)nSۺh'#|Mu?6W/_/ fw&'sn>&qL5> ULt2Z9:ܜwT֏dߨV #t<@>^EPLc`bڜt2N%_שl4i.ntq77|&Woޝ{zݷ`,u7<2wMYanhCs<^b 3]QMhW=n̎|˭ғϯ&QRZEO~ҁblt{rởϛQUzV=+͖̝B9I1G.\gk/nd^=| w_嘗IBAf2GP86$D"ec=e70(IOmm0}8^^C#:IJͬ(SK!x$y^]2G^x@Ē(A``)vQObT@DΠ:8Mr hވMP,@D<2*S S띱Vc@~{-SZMFSYt6-UGn-AK"?D K\8uЁ B a Vb[!A ED=3Er5!I:8 jx!D; qfI$*Ű[)mm:5 R-hd)z_ltw3zbv&>cXh'ՙR]6,Ӫ#2w9Qy JRi%LJ͒2e4a "F-=)'g=>:P2zE);PGI RS飶ěU`̙R1t6#c>Y%f]PeB5`b|FfWI2)\3Nl>+&ZMҌO3}Z6)KI!jE@akgHn [l^*,CR P` "2n.$UP\ٌv:K*澠v6ufjZnJfN*XCcQ(VBmعg{2ƙʊ iU3АA2 hGEN 8䌉{@Nm:VҌ ǎHP^Dlq@>! |O %$zXdhi#]86+"jn8XG!1pz ",]́(B cR1̈m:[= -bU̵OlT+.̸\pq+K$H-FDT`ϸ +PU3:\lqq_0u슇43 l^cu_u5 U?ək7&1f'S>ՌepsY+#Edp^7D8$ir`$A H5rQqI=FhևR~{ b1V*69vsgW,?Zdvu ou>}1ɵSCC~ ]^?\L2^oJZкԺlZsRmtfCF]uв}on-7wܢ祖!Uﯻ>vyqV!7س|CmqE9G.tk´~tpBfTɯ3yȮ,*{ 3 -KXs/u.W/u*i UUɒqkR>GgJYZh,Aʋ$T#5EwE,R@/l R1&*.KQ4q,18M:,QFb4jՐ:n+Pכ`Qc<=϶Ax1ڭ72\|տa'6{lR\\lYhXD^^ўb,3 # cnR"< "$ZnxL֏ &H(q(S&4FR4\3 < XByd@qpHR9);ŜSard60K{j~s $yw]Ǟ;ӕv575s?eT}CSB&26`8&.pqh1fP4(H'L/gg~^CRwEB$!A(  p#0b chV ȓyn83AYό/ )=~RDZIn f'f5^\fZ`ν&(ߢl6KNf-zu^ku쭚C{rJ^-9TuR)x}5ڶ(;j}畵{/\o_7Α#%Ǔn*o_fEγ9?팺Njfk.Njo>747ѹ_5K\}Ҽ1/짪@_˷w/&*tUŬ:ZZïxs~1~Pi]'fG(S]B R[?CkdF/Ս.1E;RB뻓=/Wb}zV]tzfC{d*vNjiB}.|ЗW@]0gc gqmN*s\p: %j4t& ڥM3y?P[ak޶=<0~W('ɱtm Υ'#M:Ejv7캻P8F{&0SĬ/I*TTҰD mQhbڇfV!:_b g0#{S\)= 7Jw᱉4c*զWnOxɔLr)Df#šo ejNVcC@X!-cU*I$?G#@o1b}F*|D۞VYX \R[s5 ]|@~ Ga~qq5fDc#9 Z<=zcn-@=h?W8ُWM\{m>|<}O.%LDioӺyiWd#C?r=i>yƼO)~>x UH/?S+sv/jOpŋnqMU}~n!.Hy`~]0, q< a^t9?O<4\ﮐ9?+ҳ4R;Nې|l|Pt[|5)PmT&}v>I#s/i s)7ƈ 4JM޵A%ٰ' ƒ8zV ԽiR\7c1\mz4i!ׇ~kn۲H~m$Y5.TfɗV1bu&qX?ͣdYX\XC:r=y<%_|&Rs;L HR $\swLh̘\fh5Ɨh,U'lBl1a-m$?껇{[d"Kqlĵ s9l<9or a\vB1\-&ĐRb)*#fL3)+Q8>„[Fa1BL?grdbպW:jZ KJR)$5t19+CE쌪jePp-=ytml%[3ȏ0QG쳳̇U,nZN{ʩ ל2beߡgs|1!d9F5]ckE<`uZ-d+Ɛ&t8yؤZ#cXbۢZcC qޞE >ay牣ۂ :#ewhbdl⹋2R,-ϰd栢Cm|c0ZK.8;nD)S鍡Jhs(a I1:±e6bk/%c[ɺj6 CK䭌nhJMCG^=U z8Ȣ#Һ>P [ bAE &mzgAydW뒱HlM\7o|:x%dْ IBf6ύ6A $+kƆ k=wiPB]"vBP!ՠw^KC`B2PhSz )]X!ᐲ`3'd F Lu>:\AZHx"fGªaB$ro@T4ТA 3D-Č.,U r 6P _ kJo7TѦb8 Et0GQF[4] \!Ҫ#K&"y -d@B[gZGRACmw Jq&-.^^UuAT$E̾<#JOE:ABb|P[[%"qByUֻ hѠ ;qLj*V&{L!8M^23S,NQ,KӌAfǘ*INipr 3M g>;5;oUD="?uwmIf$AB$ 9b:p<`"@[0mJJ9U d`>MzC*]%40#(~@_I0/)ڽXsv*J-Ah_1S;6@B`exJkljwX- 9~V b"UPkV9+tX;WHWY g&H jFkTJѰd8U{OQ OQ!,c(u qLCu\ .: #FNê>[.#WA$DؘzXLDPVx:rB. 6KB65> 4xx;j)Dd"!c ^AoOճ4n Ķ?7էC>|BsÛ7{?ߡ&q݆61(An̾xyluGK̟]wO9lOUG!@a*!`n!3( I 1;G) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $Ћ#$̇C= +%* I oKPHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHIKH . 4es($д6NM+%) I J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) I &"wH$k@6gOqJD(KVI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %^ G9J-zs25W}?\@o뮽'\!Kq>p w0*ϧ^oiN=^sO'/gG׿|9SW~u%(N:]<9{O&4-Dut dE <6FPySMy|}u1ۍuHe z30з| ]oGW|ٷ@F0~v1.]CSbDei3Ke%Qlq=տZxPnʑz A89H5%]2.0aJMlf(0h nmDY{VWa}{oZݵtŧ׏vaΟFB!mZg ÿn.}@_FWGjS:5|k.F z.q>l7*7 t#x3|3[՜f p%flY1&T* cLH4@/b H 4Bäb|4 uwx bfwK g~/]5XjM82նbӲܹҫ&7r`: ?Yk>L.|+sEDV\'3>\87j} \$~::lo9sD)`Ģ"_;rp5]?_ImdKK!S##\"0$Cs U-(&t7mM9aΦHƣldA&aanh t uaN^3zİ DW ^V_y]M.xЯu志p:7Wc@]Ux>܅T%O IQm/UvyΒɖ5-Iy_^{2Re^*/CQZϓTסuJ*ӲQkCzC^Poo߼Nh$cCVyY0I`6ӛOVwY4ɏpj_PNoEX:|}@Q]gu}ڥc0ISmXZw](%[_}n~+/u4DרU7lzU+zKZ4`]dE4`U/ ~Q7x*CUK/jt(![Ny$u$ 쐪x̍a))e@MKEqEdLK[BB"<:Դ520G--ȁB@зu=抭wEKKGlXP)Wq/fBcq1r` ^Y93ڭnc[֓ B-ge+'ԏe:N.F  4ZW u8mִZQ|g&,Ijnx |?ng{*ip2#Ք:?^]'pz#@yWКXHq[JƖa$tW !P$`>]iMXcu+-{kZw*.K0a)PJJMH *|ċh;GJČh;9* ׃E.%}W +vn4~>=uײ7w>Y2JbT-eRca5hAh*XʜBHQr|H[wQfC7f]ڬ?"($l^0C="Q$Fdg $'Zh9+kO:O^0?ѴF"¥ {2%QkgNY8= )#+&aoXQ #N)cn#H跽!i8H B@)*2@ *{$EeG:GovH&*kej }[ۛ _),'|:VuѤ|UfHii׻07yz=+*AHP8Coi+EC"l_)IP/E ^9B') TUjpxhVh-6`MHQtSzr5ܜ\֭jCpy~8(+(|>2>օɾ _zW"uWK/3J(ɠ:q{!gߚ05Ή 8>l7W2uiLkmyStr~z"!2RX%#Q?O'k`B!i$7!h06娖Oi)>*_f ]--Mӵ}$7q)A|*fi@|8κ*F)>Qs_:|[xⷦ~<_@O?^^ ޳O $6L¯w!@Ử,0lhCssW/m S04[+}7?r DϯJ_hx\ꛟt85;LNٰ͇y깶oRْsC_E>1 s9PuhA&p˿p~r߼U?3">`g${9`Ҥ {B]3DXiSl 2B$=a&z|=F8H6y}ॎ@^#:p~[*HD X.Qu:Y.u0^ݑ}7:!OОf[ F(l"X Bag!κhIL)iSZHit*y4N9*>f岑[)).@?S:μ2 YSTybU2}R6ZK[v#ijczf8F0K4S߻7 @,EǸ.8πx ,.f`rHR F CC#FzygN)d1!yj{3jg>si*UTk%i];fq ,w7̺X`d(aϓd@fz^CN9ÐxsS#T]/jWℰSDt4r`HGdMGq&rb3`cXenjkm@PfTetrD`6UtdٽgPd<M&7'IvO :Em٢WtfG2JʮYjAח< 3CTA:0eQG5 Zvc=vjZ;aKof!#p<(hEIj1҅ XDLBU9񢃟?WG1 bRŗ.py&9r S+R[0EE*juy9z2[O+㼺c׾._خAv. [>l;HHMqmi*oNy3(\o<_3,:2 /D̊B9N;uO=u=B ;2d({pi9$J8 Xr@ԓ:5*QJQ3\@FXF-LG dJaj3jX$"LqhПlUde#^c"X{*,sA*8 İoD^aj"rBpO0CQ50$5I~ <!/ qfaHB9s ˽̼u獚5kXƅqЕ-V6>9(`vԉ*7dXzmiGR.g;zJ7j;[D{ԙq$LK왔\)2PrȧBOz$|t8o%Ud>ZRPGI*R5S飶ěU`̙ rKl얌J6Y3*,T,WY8(m2Mt=I7nw KZ4(`0JlR{"9 f,R.csRq ߝB(ؼJ'U42XZBITH'IXh/#vHHL 'r;"gĶ/"qǡR[gںA l&HuÈ"\[ slIu|1TVyY1 !fE;$@Y$pRUd#Ju&wG6rvZOTDJDJĶN"vqpHþȈvY $zXdxi#]8 mVFHHqbKFb3[XA`jḁb%b6rvKڂ\vs3*qf;m,ኒ QZgJ@[Ѝ!dqW)kAXVlb+EfC!,aDyQj29, "^Huj8/wdԱss˜0# cFM@DDM}KMcQFqn})9R1s`ήvе3au{ u>̾ r+i]aqeuRU'@bx@ZOa9jG6ەAn_ͺChY@˙}on-ﭞwwۣ祖!t9{9l϶i9Ҭԫl^-)VRQqwmK_&xCX;^XcF#yV [=Igߑ굍!)B$ԲA87iFELҨsCM`E;%f@ݝp~ګ,Z:{8$$TT԰@p ўc(X$7AK/7lC  &ze`G $s8LERB)ISV< Xyd@>eh 8$) `DrC@,s< .0=B;~&N6ظiM#o-(4XqSOKglS9T j +NyQ1(*h5r Ry;hdw34+0 qFU G,"{)0>b:=o=]y\\Lz(1_:!~9/g)fc lKN{ꗣo@p?CamK2A(E"`*/)gϏWK,Rg.8c|fڿyxYjN!7 騊~_)zm@Cl8i o[Bߪ~*OLL`q(z*s<yӯ'5A?Ͱd@ou-^o _ [  ;!bA.WꚴgҬS$qx>kmFk(?]rp8U x_ٳy@߷~us%>akj>퇆v'5д_ߎg/<љ9x}k;.GOO{?c+4_?~Sh'|2>ZxjG|ryizC_1=trL}*h4{VtgwtuƝ RBubIK#Ff|z8<9d6J~:bw?ujoJ|yW<1(';&1xdD]9pw2AfIjRۼ%hD cgyZdk6i?A7&KSܚw:w _yr NC$/jY$~OpGK$+BRL+\ 8==\]>y%Ҥ X;zWJ5׽0yփ,|9/Bīն7"mcRIs^2Abh3.ӃqT8]rK- @ HBfV PX_k֪'ZͷOz0̕ITi)Ż?^gA! R7,֓E:|X+o,Ha]D2^\HJQk%m#A C""FHly}3d`O+'1#bXt% פLʥKz";'B~!jlvTc+Z1"xJl980c:"/G{0%2jiA5l}KKxX=' In9Eb+EB u"^E6T.=_BHu5vVUt"DmC]mt~n^ lae]=-XJO)aq0Q&&}Lbruwv>z:oD.|hn+Z;-ilgj&^r"9S3G[iR2f9tV2Si!o/Iگ 9$$V.€M'e;Q!(R9",Ԃ Zj6rkvz>>f07h&K#C{z7_bEtj6FE- [0݌0# TiA!u0jtDѪF& L[K 혅SYJyYKPGw59e $)G(&:I@KVz&4i@*g @%>bFR@ل/p(46˫Xœ :q`UD&gp"JFzM7{,AZ/*1u>XN5c+'$E'!8hRXY 8 P_4 NyX F;t%eV purR`yp,qԨWuCJ5@΅໠\>ږY]J@QjWr~<\T6J!^kR>gI1邋npBog+ow叠Q(Q 0 * p"*E:TQO ȵ[,, C!fJ`6@ >^5X@U6&9THTQiWmk9N^i6|{?lKuklյc>*XGɭ9#)Hn@8L/^T #*)dF Б|,SldE3ڢ&O=ndٯOwMvݧpV<\ʛgW_'a^`m E_|> bҊcP%R_CP~{[t0Dq$LK왔ҍg^:;oz$|t8o҄P2zEX4`Mĥڒ WȌ ^`A Ɯ AD6]%fmPeBaλOI3^]ׇE/Ʈ߽J?{wR{"93s) 1DK@ l1 @zktRE#Kv`j-$& fE e) Ӂi]Ïm`b jgmQ[gFmݡvVL L)6Q *2(hdk+AI6x&FAXF1q mVD #$$8pFBb%#1)DX&U0 cu?xpq'_g6.qf\.v%\Q$ Y BIl#AY*gGZ *ҙ}.6/x;Cvx[ӘnYkb7( LJ~LYcACzx) {_3(5(Bj]4uvAgu48>(zpptq'9;'m2W܍ZrԶwv02(PRfrA}"ƒ'M 2@RC~ 6 M x|l0xǥ&"D"KR.*;wT>J؛ҮLv̮Y7:aNJCwbiIvj|d.&tylFWFL:yϮ~Մ025ho#6v6|ceFpb:|{ބ7c$k&nJv͛лWf{MK.?}y}z_k r\JhayyKTzD`-%6RyLQG̢l4ﺀͮ<LxmcD ,(4q,18M>(#I5sC uPE;pQcI;n~;$YI Z l(.p!3SeFa"7 %cv62#Å 9J{&?{ƭ};]!94IHуi`iF\KNp,m9٠]j9rÈM+dJ4C0+R]J.H! @o6WqLH>o` ]ZvPs嬃V0, Ut:IpD%U&5OAIXzeX2v3:N4N|ub^CrDJR!X#DvRK7 b,ιT*Jﱓ w5.Y~U>. \qfuBfD]u~zK[D,Ui;\vY{]jqH7.7? x:szCC]xgm#M?)u=Y3]q],qZOjצdtzJݠ ]kw f͊([MwYۜf\|O} 3˛=̦n2ۧƇV8{yާ:k[qtduYzo^ۗ+~xI{5ϧ߇3n*h./l4`qtĝw<)\~yNRgotv?EsiH^ѧWd|2~Mzv^N.޾Ex"crQ~9&Ǟׅa6 j ՙ=3] Q=YOEB|G*ѵU&_m򟍞Ugn|v4:}Id6&駱6qxX?Bz&fLI)n9G:_g3ip?k$b~* QdfQyl'O,(waw^;wV 6)h[{l;'7eUwhw=ʤs),{<'1AJGR.3IqǸ:O+n)mvZ:k<]N,x#f>rdޫ$3Q3iR !9(:Eܛh09k:׊++0j&移FmreŒvq˪g}IQ*WI]|g5+>{fVNn*;B~b*HU>ZQ~mv5ZV3@[~ 3)Rѐx!*qK *@ң w]nAճ.Y`-%23$e% i͐ %SwDe"V.=)4PWߐU3 ZOYot']4K|R:ƆJInAqƩ++*'\t56@=eG־vF 5i2a9!(H#9 +낕'-lVV{=^ IDDs2 M"9ZE*&<]kl}I R<=p}qy\Y[)*ߨ2Y1G1?0b#@ Dus#ڷ_r+ ׫i#tH>*JE:ʡ˕+k0 @]z};t2S[>;k1 _j%"#@ Ain1(s2n58 H#.u3V91%DVr V"r) 2kۡvw:.r60~h^xt~fU )ru֔׌OH*$E=&EM LH"vS0T(1ܥDE Z9)P}8=ݔMcKעfGu3ű ëOWRF=L&1Zay\ϖ13FYBx* WIJmLzpLpKBlxgִG|VMZQ` .HG# P5D,Th.Y2LGB!.ԮP:"'" e6;F4LE:OSoIM y?:έ˹]1ڿ-{`'T=nDc75wX4fbG|>' Ӝ7*ͽkXIZf9-Gg<|z+bchz?WjMմ&~>?1-8:|7w/oo_~wH{᫗߽yA;?h&)&7 އ+]ϛw-yaisԳ _<- ߫V.~Xl5IC<͔1/CW9=K(^&6?ǽ0{DMՏ"IE f#>c&ry)E&ʿ7ЏȟAg$ ,wVkȜ5Sh5Ϥ6)DB9(&Hֆ̫j/xY4} sJ9+RNrк8 G2!PC*tښتa n>G;_GqU4KFHw7܀9 /h`u%e;/*ׁy}c%f>Zb{c5E.ERet.qHG0c>zٹ~Aˤ>G4PdM/U~*([4rTWC 0[ym ovhHиo"VR:T, UNrب,ez,O1Od,ZIkLи⾂e< +G0C^I@f_ߕi[M(Az#>DM/Jx.y zB]o0oeF?QącYͥmHx !~nA%&9 U1* +_y60.,;arpqi-;SI壔9FN%b@1euT[= Z€ʐcr+A̅He ̶|v9:,y*j>_= bz8#?{۸e a, )! ؝dv'tg?LһcM˒[t{dJlJ%b&ITTսNݺJjy}ZhkSs;*):}E멕x;a_gi cG!xm]kT-s<.ļ & 5O?{DaL~5AVafJil3*iJq35 Zvn:{,^z 2 AIHRC| 䥔.d&̷Z2߀~%(ۓԦi  >_ua/UIIL׻xq;s@5NnTO8rGS$KB&TLBrJ{y'қeD&ۭREm]Nb ,.u<eܖ,I mu|P9]˨ÎVR& Y05L C<&HSE51j2!}.{Br"[sud0Ukd#\ے&24و.Mդ]{Hѓn?1]WNh盒vCY @W}cN؃ŒfG[@ɍ%1䛟GϷo`Gr'iqvY%؝|.^Jt(tpQOĿ䄱9@WG:K~Hq/yl PX?7KUb/.ΰ*X(HrKsy2㏋K]YQn-5Jм*?-|K6~=(+~=ŌA퉺CD޷0E'S3vr56Bbc(nWkK{üL N+ؑ* vLgr*CK)" *k/#G5xM24uYP|9R mvHw2<;wteV(ҳ+K]߳{^=%b]/n|,IJcIEAz b'-XSS\NZ >n;ԸaTB -4&uc׼16Х̛nK$|7fy ֖1%$.߲*K7Gܧ-jfjWbѾR-jg(%YG5tRhl:]EY#]QD*++o ]EX5"J::Cb1}Z .mTi[7vtu 1ZDW++ڣ"ZNWe:Kx}t+"u%bNW9U_f=tRh9o:]EM;Е jSiۓ'4ՍWWRҕX6+1m ]EQWh`DYJJ}ɘ'3ܣH=XoC t%Sɴ\u%Jt"UJ0;4hz߮=m+%i ]EZ0׈;:C"0&ZDW| hOİJI;:CB&h ]E-tQN]#]1`p(]E"\u?BtQb 5|(]`ek*• GZ7eyFЕX]hUg&^k÷Ic_$E&.vL'(u^iaql3sWz篿{>xDc~~v? .XROI &5d x7Ϣ䇒6\\%-8D˯n7[ƹf??>Ln^坟>7.lwZԡV|r|FPnDOjzT{{wpim>/(]`ybdFebDr3f #4˸ׁHzcS :%5ϷKq{ Pԇ˱IX¿O Pa9$$V8m0RS-y2Rcib )pQ~)ςFdnxwLoÁR"e g`KӂV`$)AeގF T.rsZ/6-u]-Y#zb7鲖+};xcֲŸk r9k44 >A.p^Z픓. Z} (WKM !¨U1 # +܅INk jtAqwqЮ$_/| cuɻx Z2vv2 80FK?%y%{bp20 f_zK@(S,eQD?  n@ᾖ^BwMcA\8 RΡʗ -l~֗0Kp5k:%yrS3d>)H}O0t?kyE,rBH=IRe5D4<0(2x S+Z8SL.7wU=Ѫbå^}^tP,mj"hm:mhE (R"ջb?h5MuB0A_< FR,{U(I9ې Ǒ}rF =п <@d9oף ۺ蛟GFQK+biYu6h<;h(DkހǔUʼ'wփ)DZ0kiPUQh0"J;Iu"IzX]*g8s2gRՠ ҌI2-=r 4d\̄ЩH#}JF7.5(ՔJwX| QŸb{!434`51# ^FtfX)Q; Zdʽ3ø)aYQTܤ`/6YHR}V*r%I?v {~{QQ;uӒmбeUyJ\nMlkTr6UHCSKIFD}'bӮfY al#Q"4&h8{Y-h!3?=ߕٲz;e,SMj-R޹(co:sGUB$\$ODtjc z)Z "+p2*Im -t%is}f:=`h %p &N Ѣ &eF2 ,Z#ՁHOg2Y_]G_v+q޿?ψ=f3B̽2w>Trr:mS² 1~U2sBׂzF,(F'd]a g`vh(XVG#$?g:EKbLx- UĖ\|ƣ$[<{m[ﭴƮ,,_ YaXO\L\SI}9;GT(X+LRfW&1$YY;j='RP.z*"jEs6wC,H@1*PIuoE"kFǬpuXWS?AD^ޕ|t* ^\ȼ6eI C%28C*텫"PB%;&?o$( Q.()ᵄ 5 %9Hlٕ-Ww]5穆N[.qӹW1SDZ+J(RVhå7"( @hx)f"%.ҳ ^n7eLPy4ѹ{6 i %"Z(@n>b +Ӫxqz+]zn,%ٛ0|Mԉh&:bZQl&J+C:$GN8+STߨ.1`^.cQ:OI'He^;2 ZPNd$F`t"(5,&9 B]rN);C]3lg^/==NzM,, $L[] (9!~Ozv<':AZ sAz?8ލZIS< G5!C% w$=[ ^jlSK- W%ol/ QR ׭gE'3$OXyF+<%AYCET1Qm$Mh5 jYLLjk@4 v6s;W5.: C`::7NA C 2$6I^\8 )0a Vkߝ9ө-s_?C1g [zu j^o]:f7t=ۅﺜξ$r-iȟbua"뙮 SZkOUغsjz}zC6ol:Ėvyxs糫Z[r=?2͟O6Wy_6|G_3+`JZFU9 he\2;KYu*k5(ƥsr,vwTV{A%*EGTb Z8!HE=OVɹwɹfSz.|M5/&v~N[Z~gr;w:*dַ!%k8H | #C'A78v3E-n iϙgh$|R )iÃP& 0QΥTL@q43qHq&P892;Gb@j=+ ٢[;vW^J?רG>xewl5~ns5+Pqp2BK-H- 8hEB%Ihq&9C<%ɒtO'V!Hh)BJ#Șb,J 2i72,JiN;!O'Jsp0rzg4Q8CrV1 \;>%F߆]juRJvR_LwJ }#UV(8dehQRJ[v}yDҗJxY]Y[WHHZ|T"K%'MTyR΀hoPBbv0~ȵbӵkctkcp_3~$a0=G9fO(l-~@9ѭ[Ld.'&1-wV~PZOzYli-u5fh&d/wEA&FȆ$hCn~lp9n7*˚SC1s\3?iB"54 w暡~YuCe8Ym6dRn|Z[%"P.G*7 t!-}Q=pǯu֢h)b,uQYk޼"jz?0-0ʾ1ќ~sF6(whVNx4`?|g|_ۋԓ,G.zQgo4>?OksuJ~pI^wm4j_&NƘH! -)R"Y߭&ՔHKK6;AS۩soխ'tҫw?7-dG?6)áeqCgXBhb0{n78pmpO C_<6:eYD_ 5'>|[&&_jl&.įA >xX_xΠO` f!ouaWB 9s:oUGTUFAD-gqFH>)IO 0hm^NŽAOEʒ ^xnCI  $Gg(Y &A4 pά2>ED u:O-I)c)4YWÂ# fހtZ@WX!ky0I/ IOzLVzѲ#ZH/@}HuttmmcE;7̨W6Z}PI]|9#*Aڔߒ~QzҼ ?y-E~ NqC4 wo_J^s^sPu>c]-EXr<-prߞۥ56$Q|r}f\C>q$oC!,U}Q79}҇v M Mܘ?ʭ_LO [+mZ1k`))-Ņ Uy ܒ92y?o\Wo *A&Fa?B>ay~hz\s9JE*!)@ůVЗb?Lj%8 \p }u'p!%p!0tX ުZ>?[+pE!/|{/97#%K-)wGx?0ݤ8 ޙIA=@ˤpʩ h/gG_ٖ4֘܌fr<^'ד?wz- M|՞LŘ366ͫP 3: úc?\& >~>'Gbj4@g_*uKeq~Ň2Vz ~Mυsc̱t:d[{s%`0d0J&7C yL>//oaʱhߥGw޽6Vh$c)CȤVy 0I^͔V?T3]r469V]ܧm&˳78 e)sH[frAU˂M]ăWJ!}ܚ]' nG4ꬅH()+KR 3FNBʃ[P5CaLP*eb-Ll֙ػV=ˌ.~ mFAF c# XA#(?}n6Osn9$.]?I{7|SjW;~=I%ő>h~##&Zm2 G. 83i?{PX/ٸHr/#'g'+8);Ç~\n>YV D$Esy2QTbmjoW'؄دiilJGT^ cNn')G1-N% gH&aO _|FٖڝVWO+;Ö}63Oy>M@-j oTplTh5/"J:vRhUh k ]E-AtQ⎮´kMx{UGXD{&[+EtkUDKh*#+$VmfU[㻊h{y(5JbEt9Jau|WW*#+m]i,j]\m4Qnf(3HSCxܠ* w曟ƞH[M[n% 4ZuЀP.JnM-hw4kc%-3[UH"ZNW%&%o]`%qk*ռ-th5i(9)%n]`u-t>b Pj;:BbGW{*~ruOZ{UDU*$#+N)m]`ATk*-t*twm=}{UD+X*J2y2`Bjpyk2D]JDGWGHWQMd FR"\uC "JB::BL"&u1FȶUDI*c+t%aI/aSְ_kpvjh>i|Unib@"J}ps<346fU- Zn7[Ud>4oTζ@&6qD] Vg .#^C3oR5lFlZkaתǜ`Lyk*-tJt(uGWHW(-+Ln ]ET"ZNWeΓI EtEq1#\NBWTMRˎ n ]E5 *$#+.8ƬEt EɶUD+"J1ҕ3"vpۣm蓆UD;1ҕqR*֪5tpm6^]Ewtut$O ޗQWn{`DhGWGHWs҅x B)F\(&--`mi,Q{TeKZ*#ZNNTCrêl *C+ɡyG". n}d];߿@a{Y?)|oS/h? tc(E.wTl!+żM4='| pg;9^=m̽?y_ܺʞxg]'uz̆y-C/wі_ŧv5Ȼk"#WŪhŝ_B/9437Q1ԇL)JE/`k U cJ#686HFOjRz\b1VE~r0/\Y܈5. EMOa*w?i~y8 VPAs `uO0s11)$Y~ 99xpWwÏZq0``;!6)z* -ZН'e~my). 1|5.¿g"^xcDή Ip0NΖ-܂p87o沟SPRgrq~ގ? q]m٧|ܠKJE?YDܿx\?h(ēChC3(DH2r||/)Xg8=>&9yQॠ4@)MoNlcڷctLeD/y@!r;;T{Z2z5 lILh:TS ~vS{g[CʭAx rVބir#+0)\׌ɠ| )#b#'K*hœ\HC2ߜuL\Y y2.͘)CP. :մB5lX(EF9VP]OՄ𭮮C*qw(,IŽq~:uaÂrT1^̈́4h a̓c,#1HZV{<=V)nŠ\L0Vj>y;hX+fLc"ӫҫʑu*lj哭AĂRe0T!R M-%&  NFw^oկԲK`(c[%gO99-ЛѥGD0+Cʘ .jܑ=ƥ&M]w}v9_ g;l }]+.Olz^{+ XVQLW<*(8M&1cH4{뱷z4[O] TPEҋhe!FQJ|+]B~ WVWW?AD^JV>zTLJb#HFjQ!߽PF;AZR0IC2O1\s#p'mr1^%a$y|si 2{+'!7aȄD^9gAj,LBP LΐJ{ȹ4%jMDC&H~{~8h5KO(:Lif\2b2^KRLPؘ &h|ED:υfQ ƃjǹ2D&͡VOS0$@aw1rhz0d6eJhMBEͻB.U^m4uMEc~ǛZhܮ5vbmTIj(:HYAވ 4M:॔E VK6íoEGuپm#t )eԉǸNWE?a~\8dӛvM֏GW q4&(}I J>:I#(Ž.[%9ٺntsTO/TO-I\qzB_ʣƇ5s r ПR24O<] 1 ,Ї;u. .m"͈}=נɲ @UE326F-{4*$1jhWk'qWgK5*?u:LlxvEʷͅW=bP >lmYsd0u=mί;㯁;iy'66tfY~D qd V#pCԪ>lvgϩo&XgSžKAL-'^7S /eh<DُʆsQZ@Ҋ*PT!(eϗSz>|1:Pct8h!cI}9"h oISIܶ Gb^4-J)D;]9CL$rc@mrpk1Y-FΖ}@-@~|2G>%%Ԙh"\+#:EQn"G3ASnR($ō@aDJd| Ϳh>C!Q 5PDpJOblaƏ5g,uTA˨*'WA&PqCto:8K]~B,+@>^mj {&q)\&*\.K^PɤJ`%#x2o!HE={k|t˹fS \s@cÛ/qw\xo]pǏwr'tTro y BJN=q`5%)PasFR{Aw8#E-n iϙgh$|R )iÃP& 0QɥTL@hf(Lpr4$-CsT9* $9YֳP-z@}@s˯}?߮{}ٟMQvW7XP3~_sxxNFh))-HH1i>7M1$g$Y`z)%F߆\]juVJR_LK }#UV(8dehQn3,/,S6sWHHZ|T"K%'MTyR΀h_PBb00KXt+mn6O?|_ ~<Wq [ǥF9^-rHhAIӕՓ7[ڢrK] 1cQG! ^7꿵!75]sj7*˞<)U 4'Q|ۚ f l=yK͡K=s3p8mB255 W5C-a9Ɏ)hymViz',+l1SDMfͤE]HKOwC|J8Um뢳qyGjo^6^-%?yi}?_Ԏ͛ m:E{&G}==!"ζ0{["L`[ݭ/&) 2Bk}&~|YW,m6Ȫ+[L._"ߧt@ _0,&?~zkF|$j`ۣ gI py5|ߢV~;p$9?D*3hLd8&lgM|lЏ͖G)>w:wV _]Ήgrڭ'6T)Ln͞Hr;&% u NqhR3qׇkCS*tXW LtF"ks"ZNTu*0@N.P9u*(Q=́>ut1ص,\\X0^xX3J93?~lMݺˑJb98V@De+ Bl)go"-g{SNaF12;)]r{PLD `PCPii  -km!fw ooV:y#)-H9I<ƈs"%hyΉh($ s}vaۙ4Zg _ý~ѕØN]oSٮ^<ـU Ѐ7I+ ` ŗ``w}?eY8 }yh,kXӽ KUdk%2vَ i~8~DE^•?4).9_N>< _Kj$? 1ؾ?|:h== cw*po`2$կt鯰jg]E.;:G/fq.tNB!q! 8V\]v:6Y^3po \z4_v';?t,?0n3=N.{4߼v.o)eiqml>\.R>D??'̙߶bӡyN Ry.m|D VJZNK[Y_RDfc{`mmo:x*m6{t,kVA3Kk9Bu;!bp.]HXBK>^#CӏQ|%첺y6_47(W}$nfi>~ܱ T6/]\X1Y?y,?\>!?|UݾyɍǾ񧳳ɎQBͩVćt]kv?yoe|<iTw3sML Oݭ/7OnS^=՞>enyF Tf`I6E2>h.dȊK8H#gko4.1qVRֲ}ЁYJ$|NxbgpO[\]}>j\HN#j I\uz<Ro-^A(%ZeyPygA W`: 2\V(\, ZL(sO$$)AJ5B &xI]"38kR̡[J䄔<ƸQEiL"`' ٻ,YǗ43<ZUM敱!r[| RۂK,ĬYG _(N1qdcY7OO6jC]{s VxtAbc,xp_[)6#X.Fov"KʊUV `8@8V}5/3Ÿ*bxJ뼵-S-B+_7[QdU6q)KP[ \>OU!鲅VlTq%xG(ytp)pNGجᬱ=BӾzgl̹4ś,-/$@'ܹēl>щ҅-ϒGhOޑuJef4, WdVJ=4c!fђ$_;7\P誢UtUQ>ҕrJ;> `UkP h-}rpԌ3 2L0)CW` h}+r,Si `µY~xE)Gʱ:< r16 +*\1h;]Unj?HRZ;/Mg.'OI|<7=Sv6~|4͂Nkxz=%Ol~_߼h&S!t#޻(HWHWFj+ŀvI\P誢O(8~teA Wv8CSC+*J>&q"]9Z i7 Cz]UzKJVߋ^*:f9VەݚvxNӛ޻W\L%1PmO'M='^S6á wAFhm骢z+jv4\bV*JK#] ]IО ]UFjn̾O e߲Gz"e+z8UkP h޻(HWHWJs# Vn8Z6hOo#]= ]i JرU33Xђ;]U}KYhkz%F3E]UBWQ}ҍ3HWp'weM`µ+*1 )U2 bM;ȎuFq>É+\g@|RYOY1>Hv.tOڨbn)ͱznmk _JS1Oj0'nV>ѹ)BxFTK67s0$[OĎfilrOod'Yh̛VϰgC >15A|3z)3!6Ci1FIà <tCVUEiF:DNs5$ ~p*ZC}ҍtutEqM+')`¥xW}+TLtut/@ 0i>pu3jO P7D)HWHW7]Uj0`Ek{OWYHWHW垪tU 惡z?vUQj1!սx?UV-2UnCKkfZEL4MZ)U2M|;+x?8SIA;PLP 5-?]YݟǔRRˤ ,"xQggp*npn;u[Tn eQh /JwU[()[ӛ+.Ikj%}!|bҋWbMOFZE fsa-8 -'@xhD9hK>wq5G.i t+mC uX>1&Yk>oUEnXH>\R<2yYβ5H8`fA2Dc!8Ca[IcXع;eqi ֐I!Q3-e޼FFT`$b],SnuV7LeP-"4 hIw9"orBŋW{AΚ]-h'ZJ uTN`,'FH8 f0!K+@k.GrDPEWOSejfLpE*,<:`p ܭG/Kv98l<fjwCBd#LA@Vʂ]p$( 4^ D q]h"<6 .JYީ`;̺ %( 02]IA1) `^A+虤+\#u VrC{Sawm$W~`z?$0E̗ăAz5-*"e7sE?([5bw C#i̼B~իf !ѿ&伖H lJ]MȲZPҰ*4]AJ@}h2&m236~% %]ЭЮ{1#.E&"9ih>(ؼ($0D!N&mroLѕ|.;٥3v /k*赫3[ݗ>Hh>Dfxs:pvttU2Mt%ڇAhRU! 0阊'8$;සŗ)XqAyjN$\Pd"U̫ fMʴL5c[= KH`Ge YLZ[]x $nm }[EUwd!T? y E*(@6T HVEO7r6jO]@,M$C.2*c%VXB(;KSv{;sP6DD z Igu $$5(ڽ,C),a3m>%TB;뒄 Xut kR([|PmN&$}NZl,j3 2Б5;kol:;% EЅJL T "Q8TDYUQõ-`QySBHcF8&gو^ ʉ6BbM[JBӞEw64IxTF+ j@eV o[6dҭ*x/GT^[Non$BG 4/tВ&ƐZJmmzvW&Ep4CS3AnQ%#pikJ’#iXvF< \T"nm4fSMhpUp4D X%d,DKV7eZa 7Q9M,")jrI$=ךIR`񆴴]JV߈Z }wlS7ڧzPn<|/.;\Q|AC/ʌs+XFtQ0]"])'R#+h p ]ZmLWHWň?\vDWtE(M`:@2R;jh5?/.&wGgV4hxgG-=6Z3ȇӷ1|ʍg pY5b sγ2-CوX0w:&R# 9TtdpcZ(Q xZFe`iCE0';-VNWw]wCkrc룽+}LWzك ?"ZNW20]"](#+ >+XtR0]"]{ pP+kFbPztutef觳_q1yv~6"]),,^=;:y.OŬO?b1/33Y.P]o:Kcv9N<=NH[wBwWG ӣoi6Z0?E'iEEs6/Yۋ^7lmHBk{!*eE(e/z7:76o?}B;]]T[VG̘=_˗E%ᝫ:e:w7g "ϗ5[/p_m?̆[|+9NgCnn= RIW4 btG/|]+i_CF[x||?\^mzD>xwgW!뮦o_J.s*?`Hi rlNV8Yۜgo6(p%RH'rѽ5Ǣä7}*)될U!ob772R(KJFt WDHqݞ??$KXRyOz}q:{弇>]M7?KK_n6#?,M޴7 IŹuK>juѲGtenH|gc U_ӏX `HIB[l|eXR juwntq(L.nP½81䬍K-5L9]2RT"BG/N}n]\ͱPw_fGВ%_/Ȩ Rc)?<#;scMÇGm_~&fљbd)hq7MdVNV2jw-^gcg=JXk):R\tZ,.''Rc?4z=U3)ttytEl5yr{X"/eI߷-Z8UpK#;G,ʉ)P'pKv_\=WZ&kY#*(Pl.lm{x^p{{\u}l5w5f~1*7Ὡ#.Ϻ1T̵TUTƖLuA ۘZӢUph^jjqF)FeFzPQTU<1ol!%*~{?RX/⫼ra!9CA4[ʉ돻][o#7+z[`x3I Md0hkGIKغXՒƦɌ" t3=mWVWrGMJg %s&%5-Hy@KTҨEr hY*歎"9R4M]Ըj4t.K0%7:l,48$X{! 0^ < >U["=1.vd6vzaC؟ !WU {3A؉ `psP<$jg!z/3T!$tX㘱&e$J&^ƭ YOh/DzA+Hد};hCbLU\2CdHIf̨d1$I,:&0C0? I҂Ta^w8{fo7_xuv#8&яJL7vJ'܏'8+ e?4Do-qjaDZp ,.q 0JgV-[rJ@ 腈+vrn\?e}< /OpտV3@\ G0x$\G<&7}u7˃o xjE^0-e+q?| "mm۲NW5OI< Ըi(YNm] 77f"¼ܚӲ`Ѽ6Dœ_]/4ݸ-1pGJwx2zF_wc]#Y9#]v #:Y^KuMf6x`YŢ'Bn^;&2[GlIv9WrndzV긒2"v2m`6N LQ_שQ:,:qxuA7߼o?}_ݛ 7O}E/sNH0$1X57?^xj M㭆Z^.q݆TrǸgyH;nZ$8?\^pQ6Zn&ybpAZ}&1\.ZK +^ Gv㥽+en3޺׿pN5I>BX}$h0qﴆɸ3NL^^`$H;ذߏ1qzf r3Lu%nVBtYo.q J$omtp~}H̖8C5g`FnuKYX铔+66opdtCBۨzƢt$ܙtIbVi;ϓc9(_R+mgRV tx5/gƏh]h.*$$(G׫__oǓpNf9Nc_w\Rn;ȬII M<_"33ϲwI<7|~x(tDbOt~A߻IJQ/b䫪מܣBdY2:sU6WݻxR;LN?϶uaGtnIimay}Vv᠘=XYϽ q9 խ#@}vaYCF!Q0/$IӡL|V2y`w9O"[t*z]Ha=7F$T92_Y_Y$L۝N&ܕ~b.6k'2m)ڸ=O Lq-6EY}!%TR窂Hf/l#:˦Fp3{xޣsG dR 3yuTh G3ҋ@dL6}[7BB!c1fIjs?9 'b ћdb%~4cZa`q͇D2-)eh;8r#s>'Re[_V*mFZUKHCrdsȸY)Eh&84I1[ $>'I`/ Gᶡho]l{K,- N5jOBQeb9,AHA;L A76ڔ [YO߳{' ZH9PJ )(YL& -ۨFU{ ȼ -DU` S02Cp,%g}IO֟6Aj(d#0Bˉ=>JR|R:rСԊa c0V$ژKKYZ%e *3N\dQe2bDVЋ#B !1B*rR0-hAF~Aixd 2NKc's{5{0O?:y}Oqng{z*;يw9_ð5\9{id:ggKfLa @FmMN)9nmmʒ ) "yL6gZzL U*հg< me,=>))*s_t;ImnL޾&G GlOb0J*1BqK@B =-#ˍ@yFvKtJ+32g66QC,\2|;f]r2vRI-r#vӆy,];ڮ2j{ v)fZg>73IT.2Rdc,-r3OBتx6>8L!DCXE '2ܣ|HY%ɩFkF~QX_oXo_ !9T%Nލ7S5Z>މFVi@quc`TYlVSN'SyKZBFFc2Zoc)MMH%_6KB6YFf' JmOq@J>F\6Xj;pvS@x:^[KIIûKI +dL$yd ~K\y_i{`d=~h J3وM'dΖ (Wn%`pDU`:1%0 R V8N8dP$Rh]&pXtt5r8l$:#BiCyfdNj6&祻p vPݕ;&5~1fui, SVkA+%)dT%*.e:3/R=#< 9A C^,RK \*`eG*SuGK{q^:TAʔ7„tJB$QފxpBe{ۦEj01O2j{vðUFDVjxe]Ѓe.W,CMTfY\ʊ,8όg*-˅Laa!\M{e\9@`g6f D+rJYEpSo??jwiaٯ׌8V?lt-[9b_I~.l luْ KawǂX b+kNFf-G}h%{fGb~7^9j=@9~󿗦?y?O|͇^P!Řh&~p9=}x̾y5:l3V~uy `x~^f}蓃 [? xzt:߿+/|u}FU_S;nOɲ=AoaJgEu&~M vOf.T?Wm 3]`F(>>1ʙ_tnūk\- ŗ RogmZkN7k[Yr4&,_ڻ57"Nnة7p2gu*X̶MR'' X"U"(k0qK4DUZBs.RKh|Ԭ5@m^2٥rHjDˉ p=N^&Ю yN JGkTGz^>x}sՔmȚ-sԫ U-NpQI ۥjORs}եB l)Eװ煕 "BDe  +$@~М4n~譞K KLA@sA@Kz/H0(P {ʙv;8FrCu݀n&f{نɪ;|QHv+d$FgDs9h<ʕ7\nE҆n^} ZǪZ/P?1PV=$Lo[:uOP}{+b&yx1EjQEINڦsKa=ק)&uqyzۯ >lyLc;WS',L:S)3 "8e%UJ ဨ kQģ~*<9I>2N%)Ĩ5pt%h*68ԉ:\^&᪷w: ԃ0 Փ+VΧ^+Ic1kנCK'<6EML0RgGhtFWIQ#jPk9dp8sj9DКB24CBujfXk $KQn)I 8ȓpAU@rF<崷#hMıhxRPlύP<0u)*ɈՐ&T')U1q1ʹJh`ЬQjʭ٨I69j$AR P>yQ H>h(_ OXll}uov9$ Lm`PS8PRUIL?PF28N5DHZB(е9? Qqʺu,;IUًiܲk@D:,i#Q'4M\&dREW(I`3Q֨@Z5tNFE0.*r gLȯQ GiWsT68k$JGHMJ]$O+M*=Mn| zd{j{eT}5.ztrўԑ'SG r_\ (&8۸B(Z^8l!)a B':q_qR~! IPQ$d72)frdr,W6tjY8?(qdeӔ~pϭiFoeZ$pb``Ռ:!r^9#G&M3o-[?CStWV4<,/}7{orix2&\>>bt Jo^H!^ë:H@sEbPךY ޢ(:Oiʞjp6e=s`546:ڼi$\=aD~o1fgw]]2]uyvIqC,<zƇ\~[e]?C+VtZy1/ʍShp{Ys通{5?G4 VNڏRrܦ0謫`$eiƹGώƛe%gG㑿я7 g<X?=h\7ø#\ݛW;sD r.IF(IY2u;RrX3Ugbt,^T9gܟE{BKkS !Oˋ f3/N%7LrOu~|Y97?μ9O`Q4~f7qfX%^{]4qGcI~ 5fj.?v#vh /Fû2g}m:2HBA! 8k_Hc 9yO5$A-:3P.-sjT*wC=#p _aUi|Ђէ$:&0 Oon# o/.ˆrFשk\6ś~82_btU hp'+b1Cx 0y$we(}mgv1F]o.MlN ~!?}|A ^:TLK9ؔle/^7tquUQ T.ZYRbѢ~~R,-"3Eb̩w`nxsd_a~!%lCv ǝ:Y@K;u:Ǭ;C JBb *JJݺimv_=˝$Rp]!!yM3!kN--9=`HFpܢ6C~B ЬЂ ¶]O)|ʀ2\IBWMR([DW)ph ]!ZNy*䴣+NEt+pj ]!Z WWUlLk mӦUFŮ y gXW]e5+DiC%J2MtBUFe*'Uve{bWhn:]eRttutM +hO*U #ZM futute 0` A[*tQAҕZF@*O?eX0#zTXpLby6l3fiM6Bkol5,k(lб]OMZDWXrpj ]eR52JM;:@b9b-+I{Ui[*tQvtutŁ0&ʀ o ]!\Nu[*Wn3MKճЕjn]eZCWj ]eNW!+ srpUn +;a'KGFdzqJ<: Y[>L)o!xZFU8)x!qQ:\qCp2O1~.*5n``p*W<+ G;??vz~sGEr< B"ZkAǐ,C@pr&f)ghRH]Q@C^Pg6ެFdU%JfUgTʊ/v1w!S #^U':g_߶4qA}|ߣmrMo-~I)>L8__)Lyvd.Jl9%.ގGP '*) :se! c. PmkuSۥjnܲADIZHyI2M 婀dRpV1cFL# JV{b{Fպ{Ԥurus^X"[t]F7Ir|}9zxs*ޣاnɱ+PlX(nu<Ǽ.utO( O/ơ]%GI8"-o)dոjAՀiMՀg˅VRe!(` P D0 "8e%UJ ဨ7Huģ($`%( 2\' bԚ@fHG4,h10Zԫ R_ތpn(n>͇UסN2 - !89]㵌÷ϧ~_|8Lj5h5JY Aڀ=1DK+YV[!y8+Em HBa ̩%`)DͶRmeiC$a{IoF *H SAH#T?r[Cy&rXH4|htd$Ly3 S"{>FjODžA3ʹ4ecFKp9)oa>✈o~J:/Yen )W^`+IO*օ4\(-*wވ|4rYWo:0x.5v 7%MyD'99j\{4v+f7v+ûd zԧ*MF 3v%ieZ^b im'|cUlqרK~wq jdN(oryvAj=`o0*z8M\o S[6Mrsvm~n1ngcĺm5Q>w[ߟmyO%#A5tER+ڹv(hTmJ"d:z޾vXK cS.E5L%acĬcMp;egXb6a 9òJΰO3GL{@e!mG/3Ff11p\D\%nm%ܢә!U.$Rj^;%&-se[ph˧d{sMjefz1*1{`\,;syt1[v8~ I;m-Y%P[֌ڌfۛY_i&Z|,Xvpu1ѓ6Wiη*Vֱ2t'Ŋ5RV$>Ъ7r9??TjԮJ5?fǽؿ qǟ闟߿/׏~|?Of` m,铺to7-Vޢiiဦ9o׭, Ccgyvė[n͗OuqukOc6G{ۈUr7euZչXgi6i9 ꢂN2Ct< MA4,AGuO+{x$hxBs9e#R2Y! /sA\G<Jx Yl"Yڅ\;A'GQyuH;LH2I/$|_~EBW9tk\zg.]1sB$h'.⫴,5VASԟ*(o*#Xyr&,D0h;mY*X7J!++S\#t%;(Kt\Aɂc\+52.l7e{v{w~Hm}}VhnS 2հ/fB)a. Zw{'Ozϓ/87d֤lY&h ̳8qBps!arXIJQ/bDTkσWVeQ`V!,*J]R e%ЀʀYrzI:s [&#g*]T΁Kcs=}zJGOHF;yމn?*i3EŬJ</r\.:LQHx I+PL|V2+84:εs DTֻFa3zn*jiQ|toYo x4mgv}n-:Eh+:4KkT,p2 %Zť^~RGU,. .ɳ9tc=vjEV0``Ih&^ ;f f l:Ϸ*oBv'IS=EN_t']uQIwË=^!8BЗW5xXe @&dshnJv,_w𥢝{[J{V㺲+ks9Kգ-|[ MҖ#G x-*g589 HNh-Sv"sJhg }ȨS զrI^8_]:@m?ۭ-]Kµv*˨0y\gNgJ  d`\f24nO)9}NQt^.Hoz2v:o.=  Ư,f1bZ2u,cƧ.M&ͧ<!u05`khS2yVƭ-qHeOR-h59yTu׻}p:Fך drF< +7W9r*6,D{se0c5Q~E%8(K<#'9;X*` cHLyݴښR.ڔ%%|,M.1ٜLk9W3Tmd֝XTjX,vXxT,\mzS.{'^|-͓?+`t;Oo=-KQRd̎+]4h17FPWVfi(ʞ'!) lJm Xd2v̺,!8#s2_|G0TPt<]evjw vu)q3-PIEۻ$F& H x%W/EVì1XŁ4dbH aMBR "NDd>TW~z;6Kl3N7׾UZ/䌺Ab.o{H*Pvt3+zdO2A?ͶB*PɅ}jwM |WJw{}Clߍ?1dmy+S]#Yz͢3wsTY/jT}BٽjnnByqWtXWt׸גֹSK?`vXmẺ)qyB>ǖ [m| )*M4 ttK,O)8 `s,kBFFc2ZoZh1OɗÒMIR,a M XTc[n#eMjtk}c$YI%-.(.IYo9,WdR$,ߏC Ý3GQ(HFlt8!sfF95A )_ljdJmZ]WZ.faI{0vz}#aeM6P# mWf2NFd{&C)(SDc!I DpC@,s< ~M15 :"styHC ;m*_uKޮ4?&:gaІ6?8Д j +NyQ1(*h5r Ry{< ==p|V`0@JٔIH)CL@SSZ#OIq4^I?׀#w{R5Hk2VTSOFՄGå>֯6{P˦Ej01M[E=.`H]LRO t>444Ƶc(Qֲ(G)Xn\Az$pl R BX㷇}\)g1\o0׈3*)8b[ &Hq8yss+=Uv.^Nxɗ@h~h}asPɿ0DOmIY$SzڶWՒdꬖucLJhJ3RnN'=|Y=-/)F-"g'i+TԟX~BKUOF37lQb4|z~f\ '׮5An'[: []e]دMmئ[6 . ^R[u|z`IsǯG<֦G^6 '`fك/ʔ=#<٣Y?}|݊r٣+oouÛyoAOjiMS7/UO=}RߛW H[njĢ zMIEd\C(#'-z;vd jE8XhsL0c>09LZ%)bZP:}4` xL`]껎+Z}v"A1-SiD EHs"^1os0 (j4}^>Bkmu5|ƪ'pϪ?uAS7-z7וռكAk=VNb>ЁрNJj, Knpruܞԟ~4^to.zf{LŌy $:8 ڂn8U . +tZzJUeAHXGOb]D_Fe)A@(G]+8;K4=t+{yk{ISw.f+Kuۼ[bCnԴzL(5J;&ுdI`DNdGV) h@aCNXwh8VE#FyR;r&mzƒbze9&eInzTX1E.jM-hAG*LIr0⊢`yS==e,P~q?7F:$"12g!aYCZVFntpRL1D+DdݛY̿f4 >+wH6A\b˗cߢ[cPmm؞7hcpey;[&H h@.F O> ̂HB9MJ YQ]Q4uaҡEűg'A{&rvq>uAIfmNsr^%ir88HUlN%b!+*o*8Qi%LJ`3/ ^9ADZz70z$|tkh-R"|)>>jKR#32x5@ .72fg;2U 53B UOLa6_$˸#8nbS3ʋeͯ; x<8 Drf,R.cs @Zc0)2養F!9{ZK(IItYB{CJGt`ZDr*3bgg;bįǂfǾ3Q`7P)!` 8&`RULF,2ʵ`PWguI]{G~1bOg_%0?>{XY:q%铯г *))+n^TAH~-2.>\1 k~ὓ 8̐_ DhIC}r0@*V bC+#6-/*m}PmkQ03_&e CӢUEW@߫1efcq1],Yј}^"āJs-46TLT|@C =+@lyܬ",YH;`'X#%#b4sCDk/3JzVvnVѯyoiWc&WYf ݬ»oa%Fʩ{t-suR\w.f[W1Խ3VܭcݼVq]v6sʲk=7χtau{f0w]Rf-{nd-K]G˩^~uJݥv-+5lwXynVm%j)VuVu2x`3%XY1P^D&5~VO]."͕|C'9tLxmcD ,(4q,18M}QFb4j 7ȇ!jMr#+ 8"5<[a|(Wʙc W5 AE%N[ &Hq8yss+=Uv.^Nxɗ@h~h}asPɿ0DOmNtA(E"`0x?mx:^-i(Ւb|sU} L<4 |xbVsUuPg'i+TԟX~BKUOF3K7S֢ h^Q͸2N]kݔNMDu2rY_ɚHeMjYl9]0@/73jm9-᥶4Tf0_#xMbh`UCrM&̧ 7^){Gx4ųGub6u+g>Tտ o_i$j=:b6OtlWO?sI9|o_կ?~s:{h◓%ƽJ{߅v/'ůrvϿfG٣ &{CϘ֎EbQiR[eAÍFgCu&r\٬=WiƺW;RC IO' Yc?sIeS~udFye% !pߟHO yrI`odhzt=\?|B>EW*{*&2׈l[^_20o u4s;.6p9꽯n}ᆻTt6U-R eӛ5Hk2I],}{biisGNqDb5Ώ'zCFr7Mcm %9R w4 i Q2!iSS ZBI~7kJB]|P; 0Á|c72j[DyٷX*@-%/Bb+SI  YeB<rl&`z9 c՗ *PM]n]r{PLD `!#T)gwZqFrCuu9>f}.Wr|{QHz 2RNOD39HI4cs"O}E҆CxwU'ը+Po(ީ~ӴL.[|WszK׾띔l/'uu;+01Hĝ'OKJ=1[޲b/춗GElɸhGHpFCGs/xpm\Dp-/g0E#nQJyAK (u(H{#"b&9B&c IQ/} &ijl0mc1{Qlq'cYh$ LG5Z+#pG& {PNFo_CS?M4^lo<̡9/g=7;o$gL4Zf'G1@'臣PQW5?$3_YP{mo3Qc/'u34Dysffŧu-?|ފ iT3+ Yn"F]zVMsݙz/>U̳{yOji<g8m2ih}K~d!nsSY+0kOir1>s7hh^TsNg`<1'@ɔv=fk.Niǎ fx=6-#vyT8?}~«W1g8#xѿ+e!$HP^.]Y!;W&Y }^(̊hXq<%,#oMW SuzWRZNNd)6nA28]bKK}|N.0;{_h.n\CbC:Mo/`!q(WgHpe\wķoox^^nif\2k Aj B+%"&Q8! bqMje;T+nZkqySJ"I%^GT)+h4dQIGtRH3,jnC&q\GT#2Q/]3%98^ Q웲>ԒD (UG'$::[nnv&8R +֒-XE[3Iڛ7!}̣g EjIJ[[edyC 8c#a[_c1LE6I?Te7(MnLNwy4G߿ק/sszuKq#p}xh2O/O逦O{MՏ7͡ia]3<mrf{~:V ?_wAZ -$}-%|j~1 _ߢ_2! iKBe6|WV{}$P}>J$(k5RBAEF҄^Ĥv 4p@ Gv1ye=1అeztܮ^#I0 DG)ǍSL) M2@ &Pjk:ŷ5[qdb.@V2̯uK왠y-mup>2VvvΛg]^匳qi8.Ρ?$%,x#K5/W:XI~wP}-XmX5Xi+E)sH2Ƽɱ䒣xU\ NK1g=>XZ _Ν:zfVPoGP/v_va9jj͑s.O?;xTEk 2KH•NR@x&mo=VپN G빳̧O8YE6$-q{CgVxSqW;]0f,ϟr{_wcW^M̢gWkffTbޞP rMPPMƛxʖyƵ8j.O EW$xkN;kd4Jc,VG#$~ht*Ř[Twm_KGE 8^<$5>-ɐ5K"驩uUOumXcgA`0IBVvF%jr47g@ xe菧sdk/ȨފG4(XKJ(/9M&1cH4[{4[O] TPEҋe!FQJ %i$Џx}ǸҾzO>oϯ`~X/҂_5>ZLL&٢v&uqUReɯS<_o^ir\CxO`2mPpc}()NqTImňGTK+?*[Fk撷U\bWc$ ut{Ӽհcnq>z ħSUGbЏ{tSPOWEoq w |yq|޽_/n-MЛ|>)';X/rN^vm@Hjz/xͫ?~Sٖ; pi8I>]4Rv=Uyrwde;;W̞r'ADJι+YR)xіFÃطaX%a\۱+úطVɇqoǥևb[زuz2^\!B𣁫̮:jsp\ʥU\1M7GWȰc]ev<Bn !WK&[zpe%j>#At{U싋AJM5 4&~1`tZQřXuzR\/UeSL=#Dr. 5 +_Sj$ċK5|qLWgR\qk Mu+Iq#ΞY w,2An.Np>2;TIѳS9!QZ ߪ_Ձ%σףnDZyC$ }RtHx bQ\|RH37F";*{ZBǒ'8%mlt)U5w S=U÷הkǵhl5dW8[-sjKPm5µ iRl)v4b'+"/VpH;x/}y?(*fgoeTKpEY (!lvݢ;/a32,>ljv, %=tl\֏~, 옖2x U \enUR \I#&GW0="*En=x7s[)•20~Dp!xb@2h D:\e.ۗO )d 6x+dv'gr1 K-\=E,9+jD5r׺v:%zbXN5p;v0܎[@ŷRXu V-J=R0ES9׶:%l݉r*wTR!RdK$PYa)hV[ .h=,} 'K^OԘqM{XYC%Tg=S 6jQ3L놵\*b=.M/qrRYקŮOniJRYݝ.lJd,kIIE`2* !QSlTAӚZTXBcHAic ׄHt4qau:h$-VعT#\0H,ɂPS$&qQ\f I'A&QcNC*K,y8~*RO,w'4'21q*M&|{!D휢D*&z9&Sh"BbҴ_#_.X ^V5kok M*UJ-QVFM̋aU.tMuM|9K|g 7-ϴ7ejV|9f&.Y!UA_u1Y  e7 F0QVuڵě'FAQf {{^h |l[K,-Pr)'icGituq6Fmb $-EHiR{^ZF[+*îep8xJl89NQX$Ut\To9'hxUU6F)RN4bJ (w[ %QMnʱcg%}@j=kĤOM,}c!=5&ȥ 1E.Rə)x)E`HQE@aHG%j@2>F Ko!Z橏PI9҉BhXQcgƷJ,ҡX~:dF~:p]%Z 7kYQ&'U|$^d}n n0jb+ NGi2S75b@ɔ $`Qԉ"CK2JR9+HC!D>P!ǂAV R#ocg;26g46) uX[,W,s6̖j'h14xß6(m7'- L .fJ1RL )P@9ISkA׬5O!;{;8˰ɥ`aJm#ZNJ7])v#Òy(ݘtڦa6-j8TfEM:( (hqaB# Q:jHB*CFːzўGX":,24pDօ$TGemx;rsbCAcODdYD鵈":zϣ葱Ge-Љ  C'*V@5FSBNM򠀥3.8!(jM s҆1vv=8qq\uHkIɾHEb$^;&4 \ޑ 7byW!K@hFH;`^e$k\<x yNF6?͊ ,l}ow ʋۉ*wr< fg7Xy)0zķkN9rٵ^o"2YgMwV,)n5k">Otr+ϭ77-en7Zfu<]{y!nVߌBpLϠ`\)F;*I_GTb.O- ѕ  :@$S,ă2Ll)y:鉶6Dc21r֗jRzK3@gijB`8mڇghLrW?{ԿMOٛ7?Mьhf? &?Oŋ/>Ne΅9d~=_.'.EV[m+J`۳ ݍYS{g5ڀLI_CefI$zn~rR/weŃA zxAEhLN{U,V|0 ? owQ~##ksV]?+Nl:Y+ <ɳR+%P.J !6FbrN@+/МL_j#m},5ڌ] ´Rϩ;M+ع:M J=-O㏈頟PB%] 1R'+KiݛhzBLZ/' ꑿWp>~^^u |2G[r-?i=.ǃ3wwvEtU Ve/k¿."t̊esެYZZ7ϝv# ,D~3`̀վǫz>#c6u;䂄(4)р)BB-Ƥ>[I2v9 Lɞv mHj✈$*z4:y^'(:֨b }9Wח]!lmnlڮBܮBnȆ..켭:`;R'%H2'ߘů|7-Toяơf2?~Y*lzɋg9N\9^7Xf./#Tó_sB8>N\w8ãE ]ik>I)n,Ϩdd.U~Rֳj^gK?³)ij3v>1Vy%GD}TH {ʙvVv={ƌ=KJmˋA3];9sMHz 2RNODωD=DTE҆颽)PwOhU},LoV[=L qryjEI%.ڦ%( 5{0]P.6ܷ_,A}fN6YJ0%ЙJ Q2A 8e%UJ8AT@*D$"<ՉGmDxs2A}:e%)Ĩ5ɌL:RіQjkD:.JQ#b8(|k9d<8\9\DPMރqA!ƥfXk%MQO]S8h'D;@Py io њȉc! ě/P<0wj][RX-2$V\L`NxfZ%zbIPŃ -)7֫Y5Qt,r&IA{@(ObNq4>Ad4I/rS:!Gk=ۇ=Vi6\ Á{ rCk=Xy-TRƗBJ-D"XK"Jtln%@8e}=MImy7G?Y"7Q#KZHyI2MIk"2 Y%3ILpT:jT Rg-ךRO ( XQqƄ<pTOnv-GZg b;KT n1㔌E~|\dwe<.Rk[ƍiqуcFl -{_>QLpq&QtRR-zpԃ@PQ"8 FTEL.rLPA)eCJq/}:int7]c ;"vظX%:I'&Qͨ+i)C.|Q zۊw̞rU/]6;8K(v˹͞tkMCbt BoVζk Cy5YԘE0D}7m`eʜf|B-lO_Pja]R+R RO'uHUY<4M4_> 4#֗ (*})IHl=j85+C\EҀTZcjDD+M}ߙDj$LZ{-vq_Ou42QǸָDphiJ#Ar5qiW^.6!,.^rrG\r*6K_(Y `8vxu?xNN=8*ZVʺ |kKI C)2%8CJ+#Pozzt޺e i@x-!HM<-FCE5hmunMŸ<,g\Mu{*$VxQXCoË0 48x)ff.ҳ.w L2xdd$TdqLh15W_Ī/8'>pzSBsW?1|?~?ykWݛ/qU9]Ie-o@njW뫟׿5no5qk|ـ?|_<-Ү ++ŀCJ CG?QCq`<"SJ}94 $H +TZ~O=NwmmNluy]7'X}vseXէ)mRTXr0STJ,-wLRJwL}SCdO39Ҟ% ·ⴅ\àodɴ%XJg+IŸNz2)+6O2O2;O68BZ\2:-jw>9roPTxmRq>OB=eYύyYigP IQiH 6Kk<ݵB]*0 V>x9y@UG]@V9n. KD'y"{j{Ԁ-rr0ǚ_v*$”Ù|)B*S:L`s4Nå[ixƜ$kt\`ʩ2ܻȝ "tJT@A!T nV:xӟG9" f>_. li_X{''\X:# v9l ˡ8ƣ*Z!]Btzz6{J:N #V?)dz%Im -t%is8sZ 2t yFPHڅho*ÁDVHu mm"``k⬙4v֗.vWLz؇RrkJ{}u:flyQSfw r c.Oƛxz')hxz;s*ŧ6X4(ARFJ%cgKpml_}gQNu,b^=q_6XZXBےal.V@p(cݽ6q09RA0LJaxʈ+>nA28XuhcoI|:w|m]Zm?Q .5.~\A*F)USp(EYZ!^T/~Eрӓ*}rr(*K+DURAoT\e?s5P̕P躹Rrڛ'hdՇĐVP\MҲΣ,%׽zJwW#C^" K,i+˔f }mRQF!`$M֩{nuu $GDWN%NW Ϣ+FZtR!ҕW~\b#*xAur>N"]I >xϔN1N0 wWo.iӹ;ߺ}8MNygiȧ 'o&wp]|s =q3>syc_,n`fONT W;3 4PvF:hӍ/b902:[ܮ,qN39IGuNu?:s2v~vsom_>қE]*LI<_g2oKw:NkǛ/0ꛅd eu{wDڗ:Yr.6U"JVDnƃ48 a%B4$헳%,?uqϢzgT?\]&5;Cpx(>k_q?ou7wͦ nߤʝ몃pJzs?qdF!^O& WWpϕE~rw|^ w+·p7 z:.}7S^rKф\kn=u(x; "yvO˳P9%pRUSҢLR4{AQ4Ӟ6gI 7KDBRiY^ !rmH2䚓DJ %*m!}wzlF:=ypGqaV.<-ZLE;`4rE7 7U:}r?GkA -+ז\3}%6ߪ?gao`F9"¾z; )CmKݕ2Ol DȆ`ĮÛ݇qW}%M$2X+݈V>10OkhczO^fP W>YV+Vv~n•F٣c6(^ ?4P5^Э{)_yja; vhWb?v(=Zǩ]M/5,]1`FCW 7jG:@R$-QW ]1Z&VG:D։Wv;ҏNW銣ɨ+UR{u(tute"Kf1 ]1\7vU?w(;ҕ5ژ1MQW  z (#] ]9c]{]1\kBWtxͱeRNůƜ sBA >(>ҵUXG$ E Fsњ(qCaCӇhJͫV-WW>n;EWa GC#+4\XђwbVJ9n؇zaBWgpOWһ#] ]igt#+vBc+Fž:D" #+&GCW X:t(}8ҕşثL3\3uAT;HWHWƤ7b~4wvuHWHW.XQ#+64bZ9b~ (G:@B FDW ؏nܕ^xtRБ%fV+| :JeԄw}]@Rx4,YdV[0lzvYwp)ZբBl T}g<9B4ZA� J,&xcՊvjߗJ1UB"PFS**ƅERsؘwmo,ŏ{7אr6WAۚbI$ ՒHՋxk*/|2Z`%J1$R%&u+2$뒲h1^i[GW.!|5crIJ[׌R8* %e`Xh! &ZKA( aƦfhf j$(:)C)Fa]q_{5-Hm?}kj,CRBTC-h4?y/4) oR!TT2Sa".)WQet,yrh. |Wὐ /g@X34b#i_?y]F-[@!X2hOBU |\&as 9tFVcm $GY`mtY kAޙg@X7}dCa4n Ek09GZGWHc"_Ge˕ fH¦ VCJq VhrЗ Tex (AxVAb9kOeM*\ J{+[lڕ@ƨ%gE V@`Y)dSH! Qc(+٥$ِT":˄wh\ 7d)K.!k Ki) 1TUdq!{X EW z#@PlQ R i&x-t*x$0VF dmFPJ/Xq%0$X{ RHU(NP i/@50Ϊ"I/XzXEk(PB]A؁IBnTzCkNȸBLAAXz@rHBWPgBE@%2Y,5%'AÊGaB*?%( 0*'8TRp 3CuUkU r ֐thB@gJ-7Qb8)t0GRFhБ%@^ 2!}[ ()%~GiUW6!;'|/u $a}Ysw,k L#R)nթJ_JA^h8(l,м@HuH*(_CU+cᛯyk& ԙ:ָAenv9 ~)#(*SNdv')Q/} Q`i>ws[gS֜3ՂKUд-QwF]`漄i>I0>FjAuP~LJ6` JF%I-G B˅هY<"Y8_4(X ~a-<%EHiD(2 !(C4V8 gl4,|OE# 0|IEHV/Vgx$ ې, x)ܠJrB#+QQgj"xn[QM uY'O:ٽ>9^lQy\E4Χ,},)VX(;Kb R;ȋA B-r!PGt3 u $$w(},C)9h0 6#*ZN:G!X;Gd]BxdWnXbT3``jik jQ0 tdI..x]~6 3(I@̔HRd!P?A.jDQUPբ`Qy(? >&@JY"RȡhjͥV O (! ŶysoQ$%;ͽwsF4c9IhZ@!` oF ]0A7LԆ6m-uX`gwEWY^yݦ]͖f\KI4`ʹFAiGU6SS # @)x(0N;a5xYc^kAGCVhx눹cAtʯ׆ oy0#QZz ;%&^bq4|¸(G}(5ҍ7SuhZq3mD ^b.յc:XPcjab3xFLQ}!?-1DppX"dl8A5Txՠ<2 3q ,iJY7xkI ͅlX@.f": Vwӥ>D`ؖ}UV8:!fIƒ0#)! S,8"4UΛMjB0ӵP?mzBr\.כtmYJɂV+$ED1 /9ׯ{o0FSUBTFطBk@ov~Z|؄veU]_a 2liWq>r~)SbV7l|WX!tBo:95z}^\HӀA_iYz+:z"M'݆G!,)&gXƮ[xHLl m=PgJN\!q19B6'&y5@#ta2 N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@(hgBNN &u'&'oK*a'@&?o@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N?Ȃ >''q\͞@ֆ@Kv t5cNR3cNhP6Z`N^ܡySBpa&gWL͡8"5ϘBb%9(ub;(v@Pb;(v@Pb;(v@Pb;(v@Pb;(v@Pb;(v@eFDJBq@Zb1ќv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';ZKz;9{Tny}_\]P ,w4j~R-/@@ {r0'ߺi\#Nk\t.-`\l\zlKFpE d+ku.".+RjRa.z P2u\JP 02 Bg+d."+Ri%jR+ '"]Z-*o&KmkڹE֖GiM,BW")C*2jU?p-?H]ң\hEecEu[cc[WJkk5QpW!҃+FB ׮lcV9gg騇F}%Ȫ,*4^Ez Ks}?0: *zArA0p"0&ԀS zIO; P*:HJ LFR>J HҩT1ƃ+0/l~*~lpjRTjNLj+M+ڧ3wEr=+zH~T*ƈ+cAiP0 Hn> U>u\Bnxpeig&H u> WB"V1F+g*D3Nr]6BJ$?wE*d\WI/BF"6lATDwIej.Ջ*`\[r $iU~̹uʚ Zn؉0M3@rӤ6@svLi_?qfS/&Ey|5D҉`qت W$XlpErMW֋qE*jR^`3 VFg+{dpRiW#xWGJ HT['Ep2 ]ܐvIWj2A .\`Ԧ?w*j[/3 v'nSo7HW$q5F\y 屃kɅl ZkRq5B\ai/&g3Վj}3ħTff\ uVc%{ڠE16;Xt *.]@K6̂f x]T +#@[w YrAl4p%'?ˠY2.K?0BaO,f` Z&Wx&tZw\S{P0yjU/r2 "\g[<H-qE*5jR*g3 OtErM6 UjW# eNBZ]\#E+T 0F+P>\\йZHW{qeQ.d+ *d+]."rSaq"KӝAPR*\ZRt WK:HlpjN>"60F+`%d+ ]Z']"q `k zT;Ÿ#MsHGZ:`QV<\hx& {{=5nJLc^jX1D0ʕ"\qE*-0F+r&#\тodJsqE*e\WHrHo1 {ͅ=^TԼ1\NsW(|pErW\"=q"iStW$ך\pEjCS NLj+xStE_ W$e]ZRȸ«pE) H.d:Hj t[?,s/QWc DF&ei "$>tLJ'.k~RzSߍ7v@m9~lKS lpEr +RB"q5F\)/ 0^f+|.BTO201^WdN >a~S[GWjS$W\pEjm W{ W$8;$׸\pEjIWqUBg+lU6".vTΆT*`\W>eFB HfԆAT J3F`7: Z%SϵxeArc{ҭeӭs7ͫr_-ܭ4u-R1hބ5M(RBTdPJ}]FlvyOԆv%}sfAȁZ[gwWWqvGbٶXm+l]/~}{ۗ(r~~/7}2PR"-|m[dK˛;k??ݯܯ~W|YMϝ @~0N! N<]6r/q>{S:;7POջomW{ D=ǫePg^_\ܾyg}Ý.'_/,ٽ7Ru,Evh ;*6A#Ѷ7L'Xݝl{-]_/}oױ6ԝ^*/լXN/.PMmZnVJ{b9^6/bYv >uR%mzFNl_ZmpCmI4~~jG /l??D7c\۝g7g4q~ )$|?XW^WA8x~tYX\j5i6\!8|›xubTl.>.*֫ۚ;H{y8[_#B * an@@c-OPDBJ&P;hBjf``0 }ġ_Dcct<ͷNU!Y94 CеAL mU5-hD9P{PDF75JnRG1e]D3 -JoVWΗ?[<맡~X[l+Z(?\]f'B`5>7`Znezm*X%-޵6q#ۿ2OVi$r8&u6qS$álkU1 ƴ3ql3L8h g<3/}=Gy  H 92D*( 11<0EU=dA7Jz>9x'سtRDZbCxX]ns"7Ѡz/!Q=&W`lNo? o]yä` BDv R*tuy1u]\֗Cd/`ߛOʖr| _෿՜7ZǤz$_;BCM6:+.*nm+T4߸X|Cxoėf߆= puV@ΛQ0=xk=.mDu2q31،U1qs bgFdL5Հe¤*ŋI5_z9O+0= hEo6?ŨZ:UUCw67AƮ_Cˆ7F@65Io2]c"\F$g,wc-XhVS K8~U|s=ږ(Vr*w XU6|2pp d,ë)~!G6i<`:Qv{m]SlN>X??{񺕨d>q\<݇Fg[ jUqە{s@/GZֻRťBWYt^ V;g=yv]M7?|SCZ5ɕgc;7=|6{j*gW@j64{ȃgY^:7u`/^zl>ໜuFO'Jt:+Ş_i7~%\ֺj'n֬~:7LujCww8ܤ >7 e }4 S!O6? n7տ[Pa)$dR1KґP+YnA-띴GӋf: ګp x/tɵ/z!zMF+^^\GWT9jQՖs3^o[/e5|3ۛv볚2PP-mgqp&-\+8DWů _;\z4=>;I7͐B&H?(55 yWOKNJӑR2l!scNeHȁüX"]ɂ4ɀEd\'8RBN;!Q;Rbdɟ̔1x-A<`e1IxipC{0%2jiA5ǭY}KV_yv얉)4"Q$R9J9'E-&_:oEEu֫ /4t?uq*Qǩ>vL+cv fNMOX9@Kj,J .`xpu܁5ct⹃d<sXA!ftdT IPt Rw1LP^!Sb$Aw?Q؄<`jǬQFYJy>(Gglb.day&I?O<|5:֝S"9]{tF T m01ȤQx x@FDXJ* ,c KA=FS *b)7{rD%2H6qk -fzc9%cIazT1 D.jM-XAG*K$9oK [`yu7c D)b9Ō_ќ6F:$"12g!aYZVFntpRL1D+EdݻYܿf8/!/kuZw*l r~E}ww>D"!bznL?0RG t/͛14U݊^{[oI9j:RjKpDe1;˜$kFލc Ֆm1iGVWQCZP>Czx"l ʰA2X}1b ^n ~*yԏ_ϗaJF$xʥ2%v΋5 µw^}n~Oޏ7xIu-\}Uދ6=38ѕ|t123BVܶmpt=ia =h>J^ތނU&fe@.dR3+J;a#mNKOmә=H]4X%t,qOM>Fe*͝AFGZR. #,y!D!e`Jaj3jX$"﵌FMFSޟ&ΞC[!'z͈pRl˅Z( /LnyV{ = Ej`HjXDĐq!,$ӔUX 58 ɓZ- UӛSϦ'.2|MoüWfmO:9Ѵ9pV3?N]ST{/eBVT=&8Qi%LJ͠3/ ^9ADZz7g=>PρB})nP>Ҵ1>jK&Y*Ffd j0扁@\nd&nd ̥bʌ£'[0_'xd6e71|Tmn@iF+Glf=r˹\p% V v  :A3`Pk % 6N0+Xh/#vHHLxNReFlFl;TP;v:3j{v+ Sl &mˬ"ÈGFja¸aIe"88f!gDю X$pq=rTiLf<&n KZ@=cY}tsc۬FHHqbKFbS, Lj\$#&fgeWm.)ufӒCqgEb4^YײrgmZm!6"iUB wA{ʀt[\<<̦!͌n"r۱AOCG " &=Aُ.[w^|Rx&3_oSZer'O4Y]ROҶIߛuڜm&O0ut=}X׳iJ'~Yn1ħII蒱`dQ,ݥl)Yg}z[3 ~ᓳEp !CL pnE7m:f8M#/[kMw\P}ډ)5v7^Twhp8mLx=W ųMַY+o (E|q^67~t6y<+ZWoOJ$LV l2RK.MT$!nRǞ}$VŎ>Hl$! :pDM}KMDXqF1Ź!D6V*ֳsO~}:Ǐv]g nvGm/BnNRNMvKvn1ோ)$ݾfb*]63F[~;vme;v5ܮdK}ݻpzcgm;*J#o[-T{ɕ6o]ﶢGs6YFȭJLC K)KzKx Cʔ)tiI!rJ)-On1>Ƹ)E"II?DrYy+H˜HzA8lTAsE !I%"YJmRcib v6qX?t2Oy-AcS\a˹6Z& #`4 ׁCQrq=c.)Y[JLJTu %\P+= i3`tlW4,6B陙Of*MdTaXK"2ml"B6x^MK0kӀNKʃ~6(&69`b^ɸtiI* E)S6vB6֧}Ӿ陫x[io`P?xh;@z2JJc,X,Qd B+SZ][S;+~:_ХRjr2{f/95/J ` i6` FJƺX-nZmT-%* g.7b hʎqXqTN9PI[E]!/3**0P U_muGo;#.` d{ڪLev![^y86t%tB6thesbO9 ]w(]&8jhਔz!D)C$k0E+Lmb$LlB `* WUZbJ`sIZ"IZ=3FdVŃQ zQ/P4"l<mN`|+So9΋6i^WIzd@PCT@Qzm+F bߌ㝡WF3qVK.o 6,%/|-7DήkA@].Nvjd +*HHsʛL i~(e8]Ky&D.a Xa+Vh[U'׾]pQ v uI;,ATO1"S{kͮK^Rztd1&<<_L;SǮ5M|ءҹ4ܜh,fQQfO-]h"Z=uB@8dLy'+[$]-]&i"E2aAi<L Q89/;OH˻tU/gsoaT q_<-y`dI_eU'_OM#* ^Ny*sJNBU\d{lcw{$> 4rQMRIx1쬬:W$*,EƜ-b{Ş6͒IcBeR QIιXT)V7_<=yvW?V |䃍|'/ke &jz]u\DaT>7Xu',@B.d;,%bMpSuq.-4~S܊&=.Xzx!vQal-=,,θ/F߻>q"ҖPH!TH>&'Kɀ{.I|9 Zev<\C0[rax_pz=xWW bA%ECgE5 [*vX9tw'łSRLRIŻ\I&:NDE5&N`"(@MNft.z~xcYbC?"k|2v1'0omUU8ͼG~<ɃYjF;g@  \iC 8Tc>*ocGd?B9Y8f|ٛ|JED G!r?`ldKwx]w7}y;fr })ܧg$:up.dbyӢ۷7CJb6Oz79;CFG||ѧUǦ~,i5?NBT>wWG/;! uga9=><έ] ?.'Q'4 Bjə-Uͨ͠XbF|? Y|7m_>08reJ<|Z]9Vvnlz\yn#aXjePFΧQuMyMOx8}z/x9Oկ}O~OO<9VuU].%p%sp鿽Mkino޴vsԳ ]Z0ݻRv'q>r'W _.2ƣ^ܜS}W |&WlgS bҫd qc_#0O_[Z!xsfjIA=xuEHzH`3",5U. ]6IIM`|&cg3tP$R"Cu(5n\SXJ1월X5 (Iolv˝8_^]p̖8Cokb) eX뼓3Eg.^r: 2ji;[IBi=gΟgCsCNC$s/B|P(5`^Rց6U6]6V} zt!#,!R1T %HD,;EbL~]F]hϼW\9ލye^<"S9\Q,x}w-lgKGJv 2ڛ  C-]/Ƞµ `A,,ԏSюO,>v>Mgs&t=8odp.K(DA]0 T5u[:_C5I\䍭t$ eR&rV':}c﮸WPo^'F-z\zoӯo*,X|T M"{ '|L-6.m6.m6.m6RҶƥm\ڶqiqiqiqiqiqiqiqiqi۝N6.mҶMXm6.mKn\ƥm\ƥm\ƥ5CP#5rZ#5rZ#5rZ#BfyY֪-.n=>+9t_zxfOPDq22У(,"no=eB5ck"=/K:W&eƙ"}V9EA^#+ov`o;}TZ,>?OY'{2؎_Dfh}LכRĮU:[K+nPg mY`_?#k8 AziKY(Hu*R N_%d@GL=VX${w|]z2;V_Mb轀0[razjgztꉞUNYwA $O^)U5`v'U JwY¨LTi'2vώ!l#٤UHVe+ļOUݴIG,Z>"yIzHTIxYTfG3[ Jl)߲&8'=,tBȜJ<{B W,ڰK\r.#({6޺twTgQ'F@:S6rAdLH B e/ŇЉh+űmD#e10A&4`LmxTvJ,CD( ^_uw[|-!@vM h 50PxC&ٓᐧ8$LUy["GlNKEtZ1Î|*DZVY^H׿Z==kdz k>d#AYH^FQBN";LQ.4~{V,Wj|z:9vaې֖#In]Y:?x&4hx.SV `Kʈ^ɆrRzx?w:cTPa5P褲msUE 1Q0jj,<l؜}hr{JNeZ8_ASRq*Uvכ6qTk'QI‚)S* @‹`8&0Ӹ3}uPHY"!vj[dR9V^L=M[!EDm3ǭ!ot D$@nOMMK)@ E'j ˥O"!aI)D<NŸ@ut"8BD=QxHKr2U7m'YWcbn@p:mBh19̵wl)oԍ&󡽾5M{L+0` \.uV h^2Q:0"wȐ@FU !9GrxuH2Q'*$X1$R*󥑱8W)Xjd0 M23>]/d&lcflAeA24)ŒB)&TP!seBD Ꜥ)Z5AF,A*\gs \*/ &&10L>"sRtŠRFl7%-]ltڦ0j;3L(8IG@ GFƅq I w1P)aytF5$!N !#3ZQ!ք" G(<`]H"JuTtbZϮ~ۂǁHYDdCrO {ED8j>C%6>1V("M P785I*ʃ2θ༆(*$䮇rUY$dn qAH'*,1A0:D8 "|$V 0<w O6*yLR'10&9+a &Uع! (4.U6q]׌vpya,몺!9}A͵ռafbiM^o^˪J}fkm:!ҽyW{fψkh뿆(*V:YYIP%S4J__N~s Gqc?ԏh;vv{=^<ߞQQٽ> `%3ZRUR1=bYh35sUBސ_) Y42*]r{PLD `!#T)gwZqFrC"zv]HwF3Y\w3 ǫYjyZtTk[`+ml.AQ lz< f2nSuɲlM0US'$,%L`LNYIRN HOuIJ5<߀ xbNuҢNC}$&9s9,hac`0K-&n:gAߴ1r >\\2fIRj7%/uؒIwWr'Pyc&U7kT.H Q & l^-L\hT1"' ·mAƃÕS;EgQH]P`H(]o8{e m)ϔ$0=7\ *NT^?r[_Jp5z am-2h /DͶYqYᯞq 9y2ߜKq%2ɱ+0* }w";%8kX3PӺeaX#Đ=׀+AJRYkGJph9[6Hnd TkՎSL1.5.'ZZ2$O O۴Up`r˹Ʒt?xOod %|[6GwR-O]^v7U3SIL^?zSKWQ`MԳQT>:KY,%tKT53+X|6pŕUVW \eqųټJvRVE0Y" 7~<kWI9Ic{{ښ;V!Yuv7zc+trn_`v۹P~E[wm|[{?`ul5𧭆>?7w܊ sa_;7 Eӷ_|. 5p?Mke3Qpz_s28{͊JRj_SŊW '黏W`Mk 㻟_rCJe_sޙS97yc=^W x sp^Y1iCՕ;QܫH}|#.B^_džڿku?(H/Uq;dqۄMKj$}%Uv9El~?  ڰ穅ʞvCPnBaj!^:$|At5;pbjׅPF#++c~AtՀ+ ݉X%+Հx)t<] ':B hAt5G\On)t5Њ9t(i uʳ,Fa5~1t5b@zt5PJ<ҕ%]/ܸ3DoJ+':BR.iICW[hC pEt7q{Ln(mMMtI[^] CWdBWvBmQ] ]@DWՀ+v_JrJDW+/x+Z ]|t1ҕNɒj^\1h9t(Ntut"X8g'Lp9,?{@OGytz-] 2nXh_mz+.`]pbndZfpӍ HW!F}ۘ+n9jhЖ%<՛UTK[xdž3 ܪg ǎ*!nвScklB旲.f -Dէ=Ľ;vNV,Kog'.ڞJ: ZvNѻ Zswwn L讀~Lh:(!M- TϝZx<4'fDWop\t5:h:] 1ʒ^5nWꗣ,9^ ]WW@ќI͂jVp.9t(j]]@DWpXp#-VJ]!]9 QÂQ|En碫-F] NWe8]#]yKVd-A\^vAq+%+@ f9sW^wNW%1UyUpb h~jn@y7~[>ݤ eszދÿMV8g\C*wN>́mwn`қ?dݜ=$=r!֜#}6o@oloMeG|Bm~ALaeQ/o>?{J`>} B\CC9=LəO*Am~|X$f!Y/rlJ$fBoL>Glr̞w݇DPbG!8=igˆ}C{ӾEq?it7cf`oVk$o{bLkX&c[Cώq,5B?Ԭ攅BcScVBR*&UXg-mgXISy/0v:e4ґ|g,4H/n*8nXU,*oMdLNZ-hstD+1Ah"t\́bkĹ F$FͨQ;qzNStk~-6p,r~{HXݹRe"Rmw qr 5Ԍ1"'ZpO=#PC2cИ7f(:!)RL ZՏ^ቈ&DTtMhCẲE eDc; ciԇK@#{^b$f[Z0VA O87y!iK7?W6 -^<):fDyKƜ9K,1~j"9 FVRrok*uusDm:` )'kuT#unE"NXRhѕreߢ'|N^LBpݜ` ( Bb !3k.*YBhLE>'..)PA5f4ESOB1f"SktLѬ=KE ўw եf ӄIz)HS`39fga{PQAQz< ;v۴J4A$UXu>:' 噩tp ꆣ/y ,q-YgeеlhNbb{EKx[( W2OkqmNհ֞+!*T-JpZ;wMBAIV*N!'[X]b :'؎~okpʆ"RIX9#fCP$xpP, |AIJ9CiZV2T7r &I2𼊞[PB]!w|*͐jPo]\Q?1l an cHse ȄvYA .݋bM:B6÷l : 9 & cBl  `4R(ppR 3'Ah2q_ fU-5Ttgv@Qā.(ʨq`j *yGD),d~G[m (ĩ&evW>0I]nbYUDI)bXgT0Fr4j@B%Kʄ@eR":,9&TY>"2Dbՠ hCSE؄Z}FФi!D1FvӋqicƬKDqbh(HljʐvH'"!pBDE_G<ئ3ƪ\}k71J.ǍZx7YWǽ}k h"X G1<*4FEy:Y8JVH|!1*#d`1&h ^1=swa4”A~,ZJ q[G+ڙK2fhNJ@A !+P(hPS`eݬ%˽eAKڳ8x,3Z(yL@Gf̱NEYIb&!Sr6$C VAwõj `@M›΁ +wf*C6(=< @Ph5j5aвppiP/zW&; M F9`pHzO^ 'a9nCɵ+FR,Bw<#@.`Յ1jʅU?{WmxkO }/[d((g33c;n+͇{ƙQ6k4ԡxyxxC&]jBlC`t,&].NI "S,An\ z @Q넳|[(|"!B`z- 0B [dn` @3_ _zY]KWQs\Hɓ !Z%g09=yMH+r0Y} 2`?~ dqm>]`!M@ "G#\}U:\ɇ1g ABo,fQ {SOgPlv:qa3.+f/9bIѴ[y ˮq} u2Ͳs?BgY%_Oul.[n \4e0O '}w!J N |K@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 uq\''6?N kUo@ w (;o%'90j DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@_'F pN Ϝ N ef"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r&tqLbqZYP*GN}t1AN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'8XZA};?gƸk,O_~)i> x FݫW`c\Ƹq PrKƥ}0.]kz^,Eo *BuB޸t% +zyB$CRF8ezDWH3?J#dtIzI}+m +X 3=trBR:]!JEttUXlR3+{G4jX]`ݣT;dt(&COtb+[u;?Zp=+QUwֶ27aH]by(d:%qm 7 1=p]ٻ^dz֞nq993CAOzDG`!Zv"Kvl8jx5͞M8&Ok\R=@_nTyxcT8jUbuz:X8Nc|kUFWf Rq# Dzk嫉?/7grB~/^|1<]bį{jW\JYʻ"w>V9nCRfZ>sĞ=I >8/>~o@tn&(+MN3 zQgBY3`\M&wgmo[tb?NoٮMuyO;;eڿ;f|^'?a=jZ6p]_Ur׵Ԃj5͸uGt]+zG2=DW{HW}BF}+DܮT3#]VXzDWG]\{;QCtte9]`! n{ǡf Qv4  =+\|6EWfY 4]#];6cl 0FIk%V|`!Z[u~Bn{n{Cǡ[ǡ4;fiI4ԦzU/thuBN]!] =+f+d_ j Q Ftt%t)zCWھЕmݟCk$*tDa]&J`/thՊ,(wtsOA-7tpޤѻNWRj/ f+ըX5 q٩L1,<8:#8xtZu/_uәTIL窒.չi|e\܀B넴PƇm> ]7vnlmC 0O w,TFTN6NVG8*o.cJ$/Uп? gnk]U>]ބwXf6낇~?|c<-QoɟZv+ v3Xt[>ӈom3b'h&[EV+OŅ}{{9r6/ó"28i\@wT%BhOQ\ө`P`ƾat^h.;,c bc~h %ԅWϝ.|l<Q:0Wۄ~CII7Nr[D2 P<]MY|uv3ǏE7;ԓOZkv_/+%]*[eU!,dQY+\Ko0<U%/JѬ@ZSeĸ0e2JB\, UҖ l :!|`,6!n2Gr8jð4v? ]!_|a[QZ7[ ㊾1\($¢nQtYi:' x<"P;8:ڒ:2+O]i~-7 R6enY(yRLWr:8"q]qTU@@U28#ӕEZK+"tELe T#JQ7>nTq !FeR"beqG44 ^)V))*bhcpG/DfX PF}>ndÒy7'_[L4f鼈3˂E\v%0E^s-a)Ip.`f=uލt޹zLhi& G7}{D_ksE^ܠoi~㦳ɦ7Ox󜇼9sӏƳI˓8;7C<ٿXJxZa0nz|[0aԹ-ۨÛi6ߦ~kc DsK|wa+8u zQ4fp@&7<߽o㸾w盖^$bc`Y#/4k"C* >Xhy)^6=_T|PH0zEH9ڎxqxyQ6ڀjbX|+et%l !S ]H5=UH e(sV!CV!!zT* V©x^&θ)(ŬOJNOϛ:|& P A tR%-cI%@ WD6DsNfRZeٮeQcEW C_g gR5!lYJUxn*:Q2VN0si񙒾*SJ".x773Iy#Lǟ[.{IfKSaϳ=NV_7UeWTt޵q$Bv~T{>'nQmL r_ $R4Hg [29fWuUuկXe Jg{`eN~-+.\2r*Z I[KG+Zd1t(RO](جK:tzdhX+ZzF-V /0R1ouYױ<\:Ea4 by#KftIlD#ڹ!@OZa{Aܗc3Fngl&XLU ^&&.ۛ i F g0ŃJBʪf,DeVa%tX㘱&e$J&I_ƭ YN7DzFꖑ6n!6,4X%sZ0DlfȌ @C4ɲ7efH.iYJő5|PW NR}G'u1-G~Uk?a1I?.Ü 3?NM~܏SZWNqܑ*p\->#n(&8/D]FKeaxZ<iZ4_ū''[)4٨>jEC.15hj9s⌎ԨZ߼iz6/5zٸ~jix}U1A 0ĜOq ߟk+jm8GYu~Mo#6dƑ@>aa4<6x'ZG ]g[&98`G]LrӨJЊTmNt}Y`El#>^_ѩQ:մ'l8n⌶H~7ߖo_9}S:{O_ӷo^щ'ZƒMO!Ǭ-i=VYq`ƽ>܌;ڷ\mֺ}}[t}}SWigHËY +?"1NGꬃXdt%xXAGRAFrO=)Wk Pk 8 J\5P`]3d `, YOm]HvFݶu)O 8jڙw$/R7״g)4LLvOA="YvoߪZ*UmHe?Gdy|_+v5{'|=wݫz>1Wήkϳc2RRB4 ̳uq稖z;\T}3)*h+2(0EY{z_x͝:wtl@p?:/Ïs"nNݵ(R:pPLmԋyJy{PJ cI.:LQHx It(Gq|V298|`w9O"DTֻ䇍f!TS|9wszWP!j:峃?6v14eLu?u[ JH{ՁHj/l%:˦Jp3[֣sG dR 5yuTh G3ҋ@dL6}(I'$$5 ؏E7G}U Y|*z֫O~Ҷ^U1XW>zTL+@s hzAGJR[V&g_Tu`| zG){-H!y-5S]PͲ#ls 03&:4 G]Jwm0B!Qщ{Hļg @ )[\āEdqL9뼓1͜g2plJ6" [1tFΗ2HR >l-x:1s2h.KicHvdIg1{zf3 ټ ROc L!&2t@*L6(`C0iLVaT%1C*"d]uhߊLI\}F(6ȾA|%+㶹gz394Ǵø8B{Sr ߜ ˬRT<]9r%s>(++il}#-?ucǐ%Y<2:qOVF@sI3dHNnH>} +!7:M.AV akZϣiN/aW_O)GU)|u>Owe 4TarteMབ;߁wB}-Q@)D6t@I?b2 d lo`M~)(`6":6 zDK2S.C>Rr`H%sNz kxgUdQ$|5:'1otΡC\6,I1,>J:uʨ8QqM]DIY^D1R9\1:3rfT|?m68E}wO_(asSbýk_v8ړ4\Q=IڷwvAR*ߣdaQ k3<1J݇2dd:ggK( 0tPB2jkrJ9V˴)K.H`$H]c974sJ]Kȹ[2vtwx,B%EenˢNRIZ;0\966h4m4g_Ė. "12F(1;t)AyFOe1P5:hIElJmpduYCp:G$خ"n&txvKmK^* L TlQ8I!%ELdT]#tF1Kt}JDzKM:@d^BDBu[dĦ!gΛIyA)t*id]ֆd@"gB jZ$UD!9PޱD쌜{!}Ձb]͙k|풇Eޱ\\oPV6*P2rFd%1 TD,G6dU"s:{8";cy(&w –n\9#5! 6a-Ap}3E? BEя3 $ d,]EfDLyqs!̱ڇ\DmN*er2C2aw,f# ST [8,go &CX9Dm{Ēb3K"?yqXl4LtŐNX7u/q=S3Ep|y~JX rPA:Y4ҁ/z>F磨}hϤ!Iei-vl=Uפ!Pnn?7z;>?bݶ/יefOd-ٰ_aѓdFϛ..hr[j*sVM>+U 8"uTsLâLΫڰ6Uln:-fP㗻X o۱;/ȭqu;ҿUz1u3Һj[os\vhf!Fn[=y#ݣ+-Rbs|VsϞW({;yEU Q2-E˭}nTmґ'w>$l|E7𙼁Rqe=-yZ%h d7 s^4?_h1K/~we[ߓL~~ÄÏ%r#'\qqT*`L3IQ# fuЋ+sW8zx$Fѳ8tt]mkjy,^lDB]#fr^'ڷOl|k/jtgfq0Ɠԗ6h1!\z1|$<]yq_^[Q#0(:HQhAV\@]٤3c^e+_ Go4l H.dDa4f!!6*6#J\φl N:b »(,J_Auasޠx1[d0lJI<7xA C)G Y)s%\PZ/.IYo9,W+\2w#^o^0z:WwJ)6ŗ,dA,9<`[-[-[[zܹɵ&yxbiL J*JKfnGI< z[rX_KWcl 3i]}Ê\cOF4wLf|HHKu93 7`glÇ JӽXz; (o ٟaQ*0kzK^`ZnзQhG?~x#s##A&%^zK2%e/:ԲvZ^Tu^sx5Q*1$hshGs-^Ocߝ G)U+2Rd:kZSO>'dǬ0m'F]{d  0 d|V%7]%ﶒ:M @LpVb&e8[&1gK}'j 5ZlO3`z&ѢEt-խUA+UA WHW\ mXUi[誠5骠++!mXG]`BW:xc]FBBi `%ZCWUAkdut h&*jU[CW-CRv5ҕR¾hwz.]F \+BWJe?wUP.)ttzJ+|"*[DWm ջ{>]#ë+EtU:< ZcWHW%dqd(pmk- J]NF=)7R2˺*UYnyL5H#x#J=fpa4Z9P*uX4-7iU4=EtU*p| ZM?t*(\;z=t&EtU+*puk誠^!] 9-xն-tUzt(7fZDWXt[誠Wf(-+ &WRJyUg]4K7-e,Kb ;KV.yM>/6/zVgd_9c'gBy-#YpB`UĮLId&Yh]5'9q\Pb6g农oN4t{Kg ˗K%9jW4 j3pi=3y*Q0)pF$bإ`\6yl@BE.MeoEJK^.׫&+sf=bY._c6o6 vflu6L9rQMhmoFuz޻VCYN{RQ6 b)I eY*ښdy1 >Km̱,|୍%B6*ZjU"ERRGU?ts]SK*&tuoj%s3nm;O {< yx]clK[?rsx̘~V7 ? G_PavDfwֈFHvcE'Wn,wԪ NoB:YE=!%!PI=I*Caq#$UxHgtz &AKVG&ERRTxӍ2(N]wR|n{H K!A9H&JYP9g]p"1ʬLmX2Q«ѹ7t3}`3ɪ<2:#8jW"iJ[ˇ4f{^\%By tIi8 8OÏ+_zRlW7+J1Ni?jӜ{p+5y\U;[kX)fL.NKj)U5V1zyԹ罌{ 3kxO`G;lCq,dJb@xeEguʋ%nuhVwc~g²w(;r&ISy'Ds<Q^<N9p/ap 2Cʂ"j+˭xrF$f"Uh:mn VweՍ O{qr9A/vUX=mBKϻ/&^ۗY」 6E%6,pj/CqƦ+]"{0xgU)kRggO&pcZs.;U5܃t~Pg}lh's_{K=>¦+L)*畚,p2[E #}L6`=b!|YÙ e4YYYj=Z\J2H>:QfQ 2ZEjxLQkRҠnT{ )B5k]4Kf`*CVq`YcL{q4x<=C#c\KR&s_< !;*$$&\cPyݠjdnT.pH2B5J&jR𙒠U+v$7UIo."8Ի30ilѥG"8EpZKuC/(v͟cWkMhc:'(hR"f=S+G| eThD;~IZޜ\Zlǟճ敘]szHv&̞LoʄG8X<kddD m xK ғ${v*!&oo^lVXVۚMٖ++L&7"#E Qzu8M(h9D7U>ix`̥F76w/L*&.ͣg9Z i3ܰnVfÂqAGhy>MbW,"SoZ7bA84,2F3L.$KyVF X6 P2Uİe!3*vá_ J;mP1p./rmЌ:kfc,bg{Dr %iT J4/^!h0:Rg]SNh?{6 vg~ Mpl07k%${ ݯdA?iGv.6~]U] ’˛0I J0br:Ù~?;~?t#X5WZx'i>bLxR`żOƣDN W>ec3]x9ÕyxMVEDs?{jWg{\ 7}H*L`6EeuU_*뼚 y ^ҐHR gC9u/ ="W 閻B䠍 e3BDN H(ǏJ4(e%R႐8tVSQ6G0CxDk|z`x<`)h[?'Q3gb>Q4AX^ M` @ΥTL@q43qHq&992;Gb@j}קsK.gp ٺ7O3B'+u>,ݲ<~ag<"1$#܂r܀\$$\Ҙ4#Z3,IDxtziy@I"D=)ƢԠ-Kp#)Ct< O׭uw.7F'<$gSΑ\kTmȡOV3,iRθ~ S(WzVhٛ]Y[ E]kvCw=0-H^37.v7g$)sQDGMROSPMY( {1a0~Y'7h? >n_Q LJ  FDK41/FOx~6ug6p^.~(2z`\:~!~>NmJCsζԸ!UX]BrU$r#/88{ŵe4Zl1i:9(FUʷqzbGJpD6~Oqn_S<+bP*vJ#1q (qb% "][I9-Tc\z!n6 ƿlkJ?v W~gGiѼG GM:9j\m(\Y-vto8j6dpU 5zٻw~U^V]jp'; u;ů_xcZl&"Ř:Au JbNV8"x6DX[Ϟ%S4]4(#Ns-jӱ|'FNRuc $dHW.+ U,|Ѳty?dQQD7ޚ墏휍=oG9^9I &g֩:ާ g."ѕ7E6?ĕVo<){Jkd$A?i;|g'Br6xs QBWuQJWZ_;6Ϯl{84\n>0In8^h tm͘1zྑJȞ\1Boಢg"q6p[(|l]ԍu|T捵1Har.K&!-ABLph@;B-9@oĤ>[/!FPGh8|9b.glgj✈$*zA yN JGkTz^^>kkq\{gÞk~Id; 63+l >(X>fLkm3ݘÑRbuq8V@%DeK +3B>3CQ_8j~eL1%1_\11PZBx:KP#T)gʊ0 }CDo.0.f#<]b$e)''QќxN$19A5`3c$ s}_wՔehgߑS}xLYOZouoǖk{'saKio gb10nEvIzmg/ ~7A|v4#WGo~Z?,LRJ q|QQ& G(Ϫ= 9KJ{1UކTPx!#"BGw6Mmk/zӲ<{>qCAKշ_VF|y_X\ xQE1;Mdz|8@MQoGuU"!|qf8-} ~:h#w~sS[s`Yz[m7z-;4KK-!P:Cei(Z2;'c~-ЮSC^dK%VG$E e>D)5BiF\P,*%~+Eut޺+iZ@lδdp$TdTB"tv-9"/C Jk&DJ CAk/<>AҲc錜2Hr2=db,ɂS$&qqdJ,@' LDIg`i[cނg4 :B0ǒ<;mĚ2:#P(EbhSbp@ :g(@,J@mrpҟTumurT[kʤ}q"prlMRHB6GQ*x>946/I2tʨ2DHGPaIWy#8dŸ@ut"8=:fȹQ޴Zx(idÓuwN:cp)'UVזs::y.MUTK%X`)*_, E(enjŀ%)HR'"z ɯz(J!$7=DBd˥tJ`U Jn)3rGdNZwB1 N9H@a) 'mV˨A3>` .q@TD&9C׉s@ee>:;Ӓcqwb}o@#۳/'r |#QN%10$ 9G{ *:+..p/xؙvc< ,2l4p.O]LYIckrЀdeUK ɶ1Nn*)cl;ȌҧDn'ѣ PNfDb>"wymN跖}"5jw Otw٫2xWy' [j)t)RVXxǛ;t2yW嗢t4 \|3<_*QڨZr[=١MZ;,6S#mgi&9Mmˆ;wTr1ҸYnqqwg*l{^rY8.q\^K.H0YQЈrph@k#Āse-8 LK-*i+`y嘒1 y"t) q3q8^[t7 6fO/{`ph%C Y'0mB`Id^˓wX{ 'KR+ JCטQI#( L{HPMy38;M5 o*=-!ۥ,d9n 7B6u&%SdFHe\Y/)Z޼6$q_!!C<+/Kb6#UC/켅mߞ`P )hEv]y O/4hmh?. fQMOMćنWj^yXљؑڰ&vPhi{"V;lm,A , \qWEZeWEJ!\YHVSF Pէ&_rp(>~Yc-X|ÏH+Gagtr?5$s;棐 %w?n[ꪮ(k}4s'VCR9n<&3Gs>ܤenQ\Z|n zm9^?j`Vv[ *PWKpvq=CY‹pu}}w#H Ye\2NCiOmY螿fWww":vi ~P`ТG' J&q f6Y`):"NUz*6z7FeAWDe_+Q#ޓ&рd :LVWMNk&q:lh1xuSZMu:Olq3{H,-F09BX`Z(*BRZaB WW/X yfzo{eҢ<\LJ+RzVL    WEJT!\IV ]I\HV0c\Uw8H+pUSrf\V ,J \i5v.Jgw W$P쪈j/,Rbu#\i H`Z\WE\nWEU;+ C]B ]q`BEZ }+R3 fbWE`"3hgH)DWZ=*~Sh{Nǭ?0D?52V???ϟpIr/<|k+L Mٷ?0}>/7l6+jڼ6+j ڼ6+jZX6+jڼZm^QW<~aydyEm^QWyEm^QWTvbGJe[}\LZq\IV88=RF;sG}OwHӽƧÝVk;| KztV7/G_ƭ.跹.q\`<՞!\{,Sqkq׎=Monwҷ+R8Kݾ7_f+Okqt6~"pçͤ8zL=W!x,kRFwCnq4hksx0^5/۲@6l .x}Xx'Ŭ; k>Lć|'BE뷶nx{a5RG jv?㢝:`M,j# M X$Lh}IfReMs/1ۻ_֣_Q6 B3wL@s'gb hCd`)"^n섍ܘ`ωEnU{!1:#83ʁSKQ&AgI0> ojm<xyE.qNÛn ߻LXӳdvy:|L|\upcR.^mNw@br1L53b$s7#rA*`8j *JD!X>L7^dvS57A";aZK$C'\fkϑxf";.Zg<%#GHv腦qJP{`~H+pXmo}k7bWxi1ýbQ2AR;!GI7IInvb `-FӦ9]lW[*UJ݈oտ94rq|=sInxفW5:w^4,O`WƾWcګo(=8TNJk'>dr`|\F`Jbg3cbD~R Rd $[*CV#[cLȂB+du&) !yof=ÕY3zޗoAH;ҷ/ewr0ܤeP=1s` @|sls7Cn 0x N#sb 7ѳ! !%IAbR(K0˓6BLd9s,˸Bi2ǚʔϝ$>{Mxi8]黾~(QbכtxlaVJD 12*3ztv2n[,dĈ"z60 .,+;M2*=OFU}KF,W\ !J\މ3?8wh0sD(/DVy/垮 Ai0PS(R 0Cʒ#2V "9#7sَs`-b}~%wӫ !!gn=!6岤j-z39 elMQ2"idVgR]te [0fo*YΕ D3bp B`PґQ\Uy ʳ~drw[6Xz߉j"u4ճ|UujJ6 $ 3(|4cOP4fj|g!R|3 Ñ_X?(bgŃZID]% ; d?F#x x/Swl*\q5:#]6זJ`! `-85l&WÚ[oaކ{VP71> \O~ RWO5w2W%fP* ) \0[hQY5}0yVܻuhNcm7Wu1 a؊+n8bn5H1ׇdV; Bm#):Gj#P050IO0i.>n]ݍ9ηܜX8um.om{MEVV 0,y$ ae2s*X$8Ǹ, ;N%֩5EgrKX8:|}_{~߼70R6~1 x  d w\Cxˡb; eλn&qdqt5nj˭ҳwʥ·iJvSЦs]sQ#[ؓ l~8VkC4[0wkQ,@T+isBos:ٴ٬_+U?zb;Б`u9`hl,b "D& eaOM,:66,!.>W(n}" ([:nzDnp!=KorN3*̯/\}ϵ3d[ hG+G~ ?ks3H%';Z?7 7U2}@KiCƌJ76"fd!.dR3+J;aOh* UfƁ\ω%QRHĨ ցĝAu$RR. #,ySCb"(L 2LwZ-"L&<6MVHKDf19;D[!C@Ȍ7|"G^qO%x.:@G!Vb[!A ED=3EB!I:8 jx!D; qfI$*Ű[)CŌ5\b)H4YJ"8ޗmO;rAmmm1,>x|K>v-.kO,j[ =Ԭɉ;xWJ8Qi%LJ͒2"a "F-=j'g=>5PρB}&B))&=NĥoW12#XP1g&HFl"ں$[T kǮiwDn%|$f{שןx4Jr%.u}uh{qD6OYw-Kh}];oZ!noyu{&{eu=Ο踩pWw%ji.ot66HF~w}QOb0YڴP eݗYOcq3َnH(~L K+eR JEdTowdT\]ϲg5^BʢI&A8D|&%(#I5r|jgiWl}olx_yXI Z :q /hOCs17. ci-?,C 9ze`G $s8LERBIfLd,sg2D O8h 8$)nb΂a8C^PBI)/=қ $yw[qMh=dlrS㛖ߔȟ JjP؀G \.VGc̠hP3O旾VCRwEB$!A(  p#0b crLVȓyn(C^'9>{'ا[J}Fi-TY&\Ԋ u Syj*eJs}6`1}ͬEU!8$a)%i.>I^≁F 1Q(ZkY#$7}@=FVqD8jkxE !BXχ}\=\ɀ`Y˷Gp<exaV^J7]}tT}C5x{Vgv<pT$fL.m~-LC]w.M7&ÇG"*a%uIŃZce%,spUb,MQ]ƫijukͿs3xӤH$a%ŮT҄t+Ǭ@ ) -t=oN$Rv>|t5項p Kχ UFk_.|Taeinpdp4QRhFrlFKpصҝb j~[.mfCD),Ԫw |[^V]jp&|T>ԝo/opq5]E^5FrJ&-YTшRh8euP9A{y)}ӿ^O4p.}|o-Nǎ~M7 j7fvH_V@^M_YGװl{pV_eïFlA#<٠eDx'42w_<܎Rr#M;Cjg.cko>mŕwN3S]~7L=Ivp߼nxva2n (κOo9w\A͚w$5q8 )ZgUgӈb41NI#]<{56ކ . ?Hۮ2 D ~o8ra@LA8gVa#|AXbԫnV˾C28 <`V X +d-"28A_G1q[ F^Z.^vmm[3@d LbW.Q]oVuߏˏgķ,%EF@Q9@ 1G)5qFKU|I@QK#9.w1Wdpʺl^Dn[zX By]( +'1Ph)c4"_ZEEliXW]X|yٞZl{1oQD''A[`B,ƁR@b@B:HzYu+$6!!OAE$1kQRk [qܥB+\K\y&J?O<|5t˻%<Ė$299c*6RxlϘxdRrނ%$f99[0pe 9ai*EBTZ1J9]C"9eƏvzf$5=* k8EР#KIr0tQ0LgC0 hIcb 4HYawXpVްVb$Sy JY$B_/Ŭu 슋M}.6>|躉n6m1OdQ0|(~h.0CW-#:ʺU9# w^9j:RjKDe:1;˜I6XwNnڒ|YzvJ i}ҳ?ig:p_N;C CޓrdWZ|) ڐH&7zI&i4WV6wh;ml.}uu>'heΘϢyF8PNzզ,kKpJ {Roouz~xEe@(>^A/ʏ(BX~}n}p5~82쫲G2v7fOnl5G G 9~-mQVJZ/l;M4- 5iy^OguK1@u臣w ~NM]_8~95#orUv6~ᥟpy'){!/JƷB>÷A~;AB8A #dySeEgXgwwpS/Tj ^rmDg= ﭢ:gk^|=xNxì|^7`+/?{mnqSunG}`XCQ=ݎlQ'87=x\ 6݀gތ/fM?[o=hB)ְ . $ή2v+Z6 "aW\®2:B(%ezJ]en ۼj]!vUAi:v ٕΤlB91aW\F®2v ή2:v ٕ!nBaW\W{CgWJȮ4^6 fu{+Hk -Շή23_#2A[Į2]W{| C]Fv4w,΅|_,NQy8[3b82V2ȝ:C}]р͹{();䔉cÈV#~FI_MT^ yPa\&A8F/,nb8[Oل C9.2\>Ӽ~ͩjU_*k(pɃ0E;Nyb׷}U=J e젔WWw/rK%Ow'yc*6('gUu*rwjwsI-36D2(n=ooΕEV'4J@fVD6.&esf;vIl;HR gsSQRvLz]~JpǎSLx^P!t:RHs굒Q@(#YLahU$qQ:J8n85a8SMDFw_~Na׻GDN2mDjZE|k'$K#-pbL2Nk|sRȂ8T-&w \Bli or˔f<>h|słnd5TIg};c}!aW[X%>[o߭.Rq+~>CV[;F?RȕZ!׬qNqiuŮֵ74 <F;AZR0IC2n ( p\ P xCx ]`gmZ]*wJ ?VXoXxFbX\0mIH3pʅε\HӢtqtI C)2%8CJ+#P5nGcj1VF:͌ >PF*eA1M%g&06?u&l1vmCmn|f[ޢ٨}J"I%^G$)+h4dQIGK)5-qA/b|BcIBe8I\ϔ I'J3^%)lYAةB !J.2(xmJuQXky]'$ ,%ћ01ޓ`Fu"0++*_6^z͕V(-u0IHGN8+STߨ.1`^(9OIGH6ŬL#$%H$J8@! >hˀ(d?3!Ns嬿|8)Usw~vJ~Z TJ˜ff/IwťɫXLs̵P?Z^?L $x_ {0z7pL5L%!_(nc;k|< >}SN~kIĪ䍣N(WkLD7|uiŦAq038!i?NVLzڶ7O﫸dRI3d'=CξԨIb|j-ʉKqt~F`}4Ex> 7酳E0@s*` {rs Αb0t;ۏ7@ȺX;#]F,";$Ê? UR|8-nuo$kGeQg\7ڹ1tG|g4>X S =Ű>$TǾT"N^YoP.|oȎο;woooϿ99;8D[n?Cshjho64jrfxq6|>ne\VO ·g޹:wR-^V2"^A̯U+*~* _h>@|d~PTת_UYQz}aGemb#FFBe FJHkPQD4$e11 lHl? QEhMζ<ӑ$~Σ)SL) MAH V`ݙN#g:8}mgh ' ^qv6[+n{SS A w|; _f C.,ʚ8{+Ԫ.e 3[ 6e&Oz7 ;.mq@NP‚7dZ,q$O'=q`dIOb査m3N;;lėG Vj)#z'ǒK`m£xH'!mT9B}25|v#j$2AJ6#:j4JXrExb +Ge>:5qsŵSzt1f򸭜m_c\e 5V/+gTQ}$AsM٦Sʅw.ڐW~)k*{s3SjC9 e3BDN HG !VޗX~}}/Ǻ`qƝa؇Y,?,W+Ee;T<ΉLk˥$޹(Q ztg ;UT=FBYZ֩xwή{קah $R t IIQ3(&x)D!'O(!oSBV4 Lh=wPNj+-uVh,c,IKPיԨ3pov^& /Pm?f0ᴫɖeVr J{;ʼ]2y4rNPPMƛxq-g2jˢ8! O@8a 9AigRiEhښ"NR^rD< Spc|WYD[IZ6ZFK,ĺLm|Į,,9 U,`|RmHlB4`=ZE2]`*54Č!ѨzzF*.P؏YscgP?3Gb ,vkxNhƞ`)yR%DJg,TA]ïp`y5`Ց%/`8itʡ`8sAH&c5ScoMBFu m'*) Z&mEc#BHlJᴦH x鸷RPX5!S4C-qIˆ 1p6%ׁV$HOp"y4+"5>XDG[ۓ H1 {:b ڼU鋮0 a+Odb֣(UmF "vNQ"SIXL)&PEĤԿFԿ[ ^V)S7{ѵ/Zn_$cf3csBг\➿8Z|B S0 wI䜻UL 5{G˵ h!>BFBhLҚ^\ m酘e 1C)Y BJNV {O_ 5r~fˋ4ـQPL iIM %(BJN2{և'08`i>N5sȇK uWْ pw:c*K\ Gbl(EbhruPHY"!Kj[4Vryݣ,{IoO.'u"p阱AHO&r)HA刓( "sRtpjlGl7;*池vٱ/j¨mzFyr3ɢ&mj@%)$)@ %C@@ $iA5ddP<*ĚQYd!qBs[b]H"JuTtbَR='30 "{""#eG8'GQ#c!1کf-P  ùOU>a@+FSBԍNM򠀥3.8!(v5TSBaD,6#j8[:&[gY/.¸H{\qqӌ׎ ͛{A8.i #j¶[\<<,6;C^~xx[zkz3ŋfj &ܨ<v'YAޏLgPQ%xdP'ND pQ&eRاC=7GɼXʻ9Ƒ9oʖ2)jr)# `īhYaB rBgsNr"9qKHbKPɘJRZbɨBkS3X#ʺK 9#z;Y~WjHYBiy|9E'L`t:*yLSOy=|?!™ ٍNC bm#FR X?G b|2b}tHfxUc笯nf95" ?h"Kڱr 7m:Fm/՗65ν~(nSغ c[7[֛9ntf`8pf!Z{=6yxCwZnMwz.s~'7l;q[ց+޹ )̖SsqJ[-W<:azKXP(_vFw Y*~eTQ`6.Z@ T7n?M]"T ztT*-=v`\).DeT2R)ؼJP<n=DfQ*/-kB`rh_~4q% w>_6V,ėJxї];@(g\K'Iu+ֈR9UdAReF==Boxwz-~z/ '?ӳ靈 ;6 /~>ij&za/Ͽ[xAm{;luqsɍ!6>n|orSzN1Tv9]f~ce9Ns0${v&OUN2'sYzEܯgSWۇVUyjgV j@ŮRmc$&nH+*Һ@3NW Z~"jǎEg]{yy1wM^[_/;W;NcL0-絽u`vK8;KMϯkgs;<O܄M~9}Aaf'1´R/{w7uެeu^.Z ;wUo?_j״z9ǎD\ƣ6P'AIWCZG&z"yBL~*}o>wGqc㧻j|wd< AA C=)!|D||r+qX'78m<8k~}Zc%UłFw|gEG1 vpؒ`6}l*xwaԯ.Ar;]LZmCjc\hH"a{\MKievhovYts1:Q~GOn~y 176Ϩ(kc-Q? uggIH4H NS W#w"jLQ\==.}S_Nͩ?wuf3羦8'DIAP'eSՃ^3Ч4ή.%`o>~Y83ADY^xX G Sr$}<##~GGnЎb1[J_ueb+SM  YeBcɑ@;P^c)b*d 4ST CGSδBwcVo:#z'1BNIOD29HI4csj rY$(m9Y|g~|܋)zK>ʙ9@{1 <8^p hyZtԁk[`kmm,AQ [s=^Gτ9fwtzjntu4S5uBBP)T VT)Q9XT'M,[# '+d.&-t80JRQk0SK6F b١T}N/Á9ƨ(z4d*8K&+;nynņ'96y-4uYr?GP&6G&&(3^{#Kl0sQURƈ ZWN8Qf. MPYh]R8pAU" ɝ ~崷Cy&rXH4l$wf(X*] ueZdIP9 D')U1q1ʹJ(`TXt\[ڸWcȑ '9r^*'o<ғHb$oQAp%8߆%u.6xlC}]lJ?}>lvƳFWx(o)*(ȐFT.zFΚoe8KUŧFyv6WlyKE~: mTm7<|,'2xnɵ*\ lhΪWC/d5=zx~0Z|E-V\j2 zj]Ml*!U2 gcu3ZV¨"~jHǑ|a%Zv{eVZ^oKzR܄]A0 Zޅ{K$tOn-Nx%mfQWyj *B\MԕR @I*ɝӨ;w )eֳ+6J+tVU\..hxծ6?Ss@iːlNwCIh3^Z#(Nyɍts2'4AsKo%ekq#7_†p~tmV^\ \ w-\ t=0˕j]`5tUrh ]O}`w*(QutYEtEAU-tU*ytUPؕ ?f{lwb6D} QZM f=`km[ЂNWHW3EtE5d Eҕ|t;Lrj.ը K$9{=OH룹O.tu?AQIW^sO2Q$6DŘ.}v5nz^{ mr?^-fM~K5U͕7>W'?VÜ{FIR4￶>KqՄ:?^m3/l֚=bk8'BR7U.eZ>0e*LNT^pW)`\ӟ*K.^߻ۦ2\S,eĵ(h!Xi*&}< NTBu,B,p 6j-Ԥ:VA{b<;xéx4DIh!9ijiiiy1*u"lyl-/;ڈS6QG'Gp(ȥ`8xTZpB# ?ȐDL4&o#wN[xRG<^>y$aG W|4Q]HDv H8l7V>}-utv.v\i~;A=4vg l2Nz3Ki/}ZLϳh8XbSIsͲYmMh9a @`裱&I#/BomTDlTDҋ~&[꺧MuM]B׿;uO,,9i>5;giл?GuGoS#|G$~v2oçHy)oA\qG="!RIH*Sf.}USj1eBTpBUZiJP).RJfR^ })D9p?w]r&7÷bݾ48 S0 JAF,*휳.8) 2 *cS$츍`L0"`dUυe)+9KQܰ18&|H7r*Pمp]8 ˺@K(.)v]'?|W~GpulԂuL$9Fc1J [VbvQu켈*hӐ^p጖S/𪞲Օ'yy5kXOa%S9i+P+S9ʁ1d5)<8ޖP"{( oK,͍P f+U%M*G!0K?,C&@pAF h>d#$/kS`|L.@ *'gH=2 <.:q(l{V^六jjgtA͖PaNp2jRBaB k60hAHaBzg0κ gJ_g $lɣ =ebm;D?XkVX+wmiRݝѵ' cl.`-JG{~4ݥҗxT U{Sglqܦkh׆B,B e WIe#0:PBD ˋp Y >ȾQse0 2,pZZ,Yƒa* bQ5j2僐[E.%b5jj8؀ޙr!TȤbDWCѲ ۤY)4cot\ɞFj>U_=SzrWzm`E,޾Egun,2F &єruʨI^`PJZ9-TXp~3"xgٝBJl,Q%iF$G!)! Шs6*%hCow5m4g렢B'"Y8qgH)ZrSSzy72K܃n?ҴD*Le ^AdYFWLo`yдʸh}Mf'6i4X&(њ#)<2abq&]RU*!H5_҅Itc-\nSM(?6q%1/q)/zgn'?\:Axɬ$X;AydwjEqQ!PnjGVBGn4J(jFFUփ %u:LMZ;SÍ_t󴌠4Qgm';ޥI~ӿMWG`I9QX)*r{UIUSYծnh8#]ʨq\(Pӽe'Z5n&f 8slܶ޶s7~ᄐ7k4x*ȕ%\^Ūbb[]`yG6'jҊeοNzt[f@R9{i%Wd Q9陦s,i}xUg0 m㣠 }Z>REc׽п<ӷo~m㷟Ox7{sh u J* _0K+7_E k-u4ڹ@W{qIHْ[O !~ c@A[oWTvliMlDa8G`"b1j}P=:PGo(YG/j`cy|6Q+΂JaרJUjx\Q M:鹝 Sʫ!:br23@Rs$12sdTgR!tRGgLt9s3zgTzvpQVu sdѝfwr.x>g މZ: VUQceU`m`XNbYk-frѣ)R{mBpCiԲTuWFzPC>4q='4AsKo%'{LȖ+V{c\6Z6Dt:&!CtRFEDt6bL4sYDZE]]q@i j]-h3L+`鉶%cmwRoORh䴝Q#uk#3X#Eueõ ӟ{Pl[^פ,ly|5E'GuNt: *% myqz@^O.&./IتPY*e=hA?oV*)&{dƏpk͒sg$= S):wYC{jwDrPm>K7kl#$b-mv|L隸Lb RBL[rKI5UKf@*xYAÙ!\vք5kp}찟r7mrZ" ;8M/FܐYhY;v]rh1Nݺ~|uY{m4B撯غUz{8#.l2Ѭ;Ė}];oS\Rlo~v{}{EwrZmt&+rsKw ljo) [lSdיˣ$p>LD;mHﴭrSZϸ}V` ?a+`TeJBK47HW{wGWA}E 'h:ˡ E<1(OL+H{ڙ bJ'-.''Bd^J+oxjTKvSǫټv~~PN\Nfy_x?y\ Y.pk\Jp6rP"E*Q HW0u~:`.)K?qR9"p3ItjxJdg8'a(0PАbN) qHs&qrrc8<՞JENNPJײ*&i 9"ߜnˏfL>\<ݭl~fve{-75p?j3}s(ŝP8 p4˄+ 8rm!I6Qxx2ly^+" 侌i@aS+p(É%K pMy[^^Owo/~:? '7glοǃ'uu^?wkuܰ E7""և>}r("XQS?RrHgymsc%*@9quO7ϵ&}'d&M%ƹ?Z7~/zMWWzR`LT ʈ*@lݐD IE.PbGdӽ?_Eop4?뮮D; Fz^~ſt?Ta8U8kmi޼\\V .v[*Dfߵ .⩜^Pjްn33aPe{|:ײ*RaܾZ ;wtw nkx=OK Jh+*DrQG׫gr+vzk8ʈ_?-zc'gl>Izs]DN>ޢ8?>uAp9nbA#o ysB_rc8+Sك+\.\hk] :7kx%ϷW_:]bכfd1XN:vЍBIs ;W[Aw=NsD]{\~/~k?ƘSc"u}@90޶^.w- G,q~)@|{sϼFJ}3Zw?W>i]_gW4}s}ZFeoFu}u-ԧ-[ђ躲2\vAE sX/&]mrsd| _2or%{+;~I178.IGӚ2HV 邍kxsQ=r4^F]nXN!pQץ/_7 ͵=Z޻5/"2֒eףlZ->NAX[. Q0X]#NA-3E<>r<#Y "N p5,ϺmԆJ6 N;.:6qaytnw ڜwźSvO|iW6 QbfwEts:l߆堬sQVIOT0ŤА vB&my9=.}3_{g!~Zwˆf[o!KpM :2:HY 8;+zYC0:tc߲.pYg'g(ti-&m[h AbRBJ61R]p*UDP ЮѲ*O`TJuq`R]5bJ9,S?4 "U55>yL"CG!PΌ͉[*@Czo'_X=_ɕyI4F 'A{,pNЌ!Rru$Ym,]PvXu5/[' ц{ѩ]'\dfo;uٮOd2ھhYb,+C[*+%Q/fpKsܞԞx9QdHgg: {{)3RPD2';.Qew +4peKOSF"Gcl 3}KP>-gGsGCJ%<[gv|:6?[nؒwS|;ZYUNDNY$.gH*zאy)T0Δ¡Cq:B<)\tBrXA4qL&ko\7 b,;Y`.1:cm=3En]H|lwGR)VSY FZy٣>gNÙ,L)]ˠX.'ia̓DyD,c"Cܫw?i&VHJքX,2%i<(D*oA vXOrPZlVsG|X#}zC_fݍaDw.}91t0:󯻪E9y[sGvw~=ۭ͏|lroD%/)}v<4w̵|?Z4+^-&[˂!X6g$9kvnm7.=F/xXzDw n/ ]Ʌncyy]'?^wzw~~7ݓsÔw\O?p{<+0:OK=޸;FgD\>8~}ާQ!6qD濽+sꎍo%`c6VɹؚuvUW͐5D*/6kF=%3s8Ro5sz=cl(Tr98,Yb ^$\;7Ҥ3>ZkwjgҊI 4SS)#xqDq %clJl>3XR͞Cgdkö{f &KbI^T.xM9-؋+ʼn|LFY/<وiNòJ;su6&m9QB^KEW!ߝE,yժαcn:[\c1HrM䎨c,hz;`X{,Z:ގi%Gx +c? |lZxކ"nC0:e盓󳷻6&Q o,~FNl:ewj[.&o\ ۓ{w;oW 6:aAV90mz-=pIG.K^k1jzQH(\`It4Q4K:9p5UWO"\[A۞ҵ6boNŋ?o`.R}ݞwGͦk=]-d[qq۳ϼb `lyiWƬ__:o/|oNRJ_{t㋗?'kp߰dեm(Gn-hɕẇò0q5??pln?_yh~ya#n #S?խHΧj&rF!Ql@t[(xm6/`~œ: {ay{<™xQ=:mmph⌱وk [Ĩ6RjPѮjc-s WEF]1]K"(Jk;<.T72j="Vly' *ʄ>QX!$|[Y5{XK 2Ѻ4(!.!;yBbGnTzCݕz- =q2F-[N( CȂAτ27m~:SmAOWP"$袱#a a1I:?%5vS}NɣD*Pgo!fT fXƯ PfD2l~a+At޾2Y(J8 Ls$eEcz^ 2 }o#@O}AA 8FCqݫ.o{m܇Q3To{HH/>HEpP46C ˒ (Ϻ=a'.CEՊؤz)Buf 7`s`Q,KӌɉfŘMZB 's[Osa| [糱VHV11#ӨۺkJ6# fDb:pĥ8`gJ_7` \̉ėIWsh#\6J Ӈi(yR@ qX(Ƣ.<ʃJF<@(E&rZȼb|Sqw8tlGƁQðh ~w60Ҭw kn>kI58)k3q*FefcW0Eȿ],vU!#;i65n(!/AܲlzBg[EtU'%\͝1ݩ"SA :87л*YMC)a0Qf13ʈ nV$of6G@=,r;1R)Vxym5èe>pnE5H{F5~ĀY/:H5E=ڔ6<ƀAS D3HFjփ*nMJ&;.ԣALJ@pJ!P?!\S.l <Tm$n.N7fߤfaw!*>@fTec EU'Xk&raSCiy/FFX|.dO 4%7aAD5>d8Uk5n)qXgs˥bU~\d 6ͥ";f j]BERC(/.XvRRAFp;Pg,D_ _ .2 7OEIHI) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $@J) $%$ &9hv\ft'.N+EIH`XI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %, D#"`t4$V,$ſND %" 4Kl=HI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI %RHI'L!\89&^1ZB& |4$݃'_NjΫ޶߾뮽7zO@9\:&p)B0K и"!/*gl{=U}UEw;ضoHMYI&wOJN"GJ-&D)?sP:!\`My2B,uVtWRW=ĕ*tʿOf+(Z‡wtw7nqϯ桬@Yog5U~z~C2\LPnss~_ܦt&$f+u*lFvͨr0%{fe$,!\`ru2/ #<@$ d+T+;fUqe!:%\`prIWV UW}ĕ5 5ՀKisU9Pgǣj*r= 2&FD%>F?w|5ۜB잡U#*DŽ =EFnwOoK" aKb4e$LZyLJ1`]z9V\&UJ7 ~Ws4^ Jȁ|-fTϳyZ~/$5X{Qf y@AidR^"fE5FN.Ps (,8`fI\C |o-kPNvcn#Fiim/lF;\mA2ιτ*.nl抐3!''H 2ܸ_dմxQZ5{+WQ(YDNl楷YpZ`^GRVR~NDZ-NpʜgD B\MЙRR%VQ8 'Cqn7sBZ{heY-EÜ7|X=;#vqf8L=H'gnwhf}lS@ |ZL3+TkHq*k3O\W%+# ju\Zp\ 4K W(Xdprq[uq*逫>JEBB+TkYq*Y-6?RpR 3rU2 5B][:Mp)Q4%\`ɓU"\ZC+Ti逫P.@[lPeW/] , v^kW pre2P+P% SRtr(lf*pC\1& +*=~ u\J9XW}g @ <\7Ww޺B !'Td]`˒e$\ZF+TJrJWA PLDuWҘW=ĕ@HW(W&jG vUqgī+0OgdZ @-}i-/TlUqe:% !UĮP<@%bUqe:!\`P ՞pf*`]Wَ'%3 \'3#Uf[ d4iܙ3M堶IjODʗm4lj'?5v&ũi89ƣQ!sC!wH0̮Ea:}hzqB3Lvkұ5AhA cJ^+lHW(WTpjO'o##\6ȥ$\Zƻ+TـKHJ &\\Nh*BvWR#<)\`%5,\)DuWW=ĕTB3@Q ;W=ĕRZt`fMW Wd+T+Iq* WZSIB \L*u\J3 WF S]oH.'$\ZP;Wv`of-IWVkP-|U23એ2+tijUȔsIQK!B+P3-ψ!>fL>UL9qԞϯqF2x/cR#qc(͉Xb#G-{k"&𐋼0*X9o6?~C~9N]b2МwY]]O{ X[ϯJ}MEp,`htix^fЫۮxG >PUg )Ϡte9]0MhF1Uz6il^(1MP ],W20>_:ZsHhίSRbBiYO&.&]2JnUeuCһ@ 9r ^Ee~kB_WYA k?Q=ft,/Rni4yT>Z QC&4!^e&7!h+`e>,>nMm+odZO1~Ֆ"q}ߩgm?Bb́Y!%#wU9^^Ul|r%ŷ ?}6w* OLr)aL{^n|"}({C%3ZXa5 Bk+6䊩:UȘ@T( 2;ʘPp=1HPXςp:EwJ TdlMdlViՌcYhZfX,*¼-2>/i!FE,L'[ޕrZ~27or'fiӒKRk&u HL:@{E[|N'7s= ɕΕ aa.#L(*6-59OgbvڭՎcm[=Xxh l3Ţ&Qj A K5)@2!`Qy!0JIx`R^tΣք" '0QpP@rpve&iVN&.]!bkH"2. DxȉvϣP0ia-fĆ D!<}bc)!&Vh %DiqS[(MyЂ|nDTQQqL&4"u[&bkr&E2[7.:[%rE:pqc5x& gQ(ssGփold BΑ`^e$X5jDZ<-W@vʻr/>h`Wk_ƙ$86+r.2El<*a=8;F؟C~ hZVg+v1'DP@*K=j`T6N$ͧشɾOgJ<׉?ocurQͧ&[O|׃O1}Ҷ?xAN40`RzV-*[?a-FEy%88fG % qtvC\}c>dQ;{/ƀO[EH3NVKle bjVwuSU+ONMԟpFoU=9%atBzH(}LaP\ mr -302[iW:ed:102,9H.h3V WTL#Ӂsٖ[fޅl6몞:a~Xڲ4x<;_k?0:$f(,W_Lw6}-C/'G0F:AuLvAiM(y`rYpE"Wh˽E؋E(cƽa #:S%A d#B6QZk9#+I' `!*2n:&e2@d2R(RN)rs㹂I쀻 Z6 aXn_+rN|8}J\նZ<]\tpMM.Z*οs#Kzlj-hNQIIpSFPAb޲Xv,`{ӎOcoz|wb^CrLJR!\#DvRK27 R,ιT*Jﱷx~>y_&A}׃:jowv@?H-J\4q[D~TXd_rSzPGiK)~ 9$/ 8V.hŮ p8?~i\E F?R8t_]aDF@*+%L0a ,xɑ]o) jӖEޮ;;nn4ЋMÑX"V Tm8/W]xt~ 7*6i\^-o춮XoTɄzWvYPqfdĘ@WWx=Ue~ fk|<vxۗ ܼ%]؎&ܫZθW+0W0;]i뙒PG _1s=RR6Q>XXe?G;8Mϧa:j6=BZzn˥\ p"2u눁R34[:n_ "D16 c3(Yfp=Ȭ-"Nl6\׺[M=ԄJ."@wu1j٘b4aetēNkkE6e]ٛ'Dd?\G\aeֹMfyV1QCB)C<&36C\ʮ[]wǡ`x]S5y#f>rdޫ$3Q3iR #9(:Eܛh09kz׊+#0j/κKk*r*cv۪WP* -n9ZLZ ε0Z4ˆMRelN.TNA$* ԧS ɺB`R]rb'YO@jtT6d8y`JG.*@ң C轏@.aXUe=}F`-,̐J$Ӛ! %#uQ$A+|^Zl7$w?Ђ{C'\3N4H,pۉjQ[9ɘdervXY庲q"mGs|xnG@uA[%C}g;ܲcA3%@  .!+f D)|4u vA*$p7%]& B*阄>-m+mMGB̕5S+;!̵5"Ɉ4ִB|B[~\n'y#糉%iuFG w."s%D^#ȋB'AӤ)-+I), R@#Eg%2V 2Y*1bD *dlOkZeoY^* e2ay:# PXr\Zt @6LrOR e&K2!C|VX:cȽfQ2A(9 AWN`:Σ %XJ3"WoV{N dKދ&ϗ66 mo^'4Ʉ(W&FU  ٸ N(SŃUIT Yȉ*B\*YN8[p\ 5{wtט>Oz*kpb,J/o$Z^hIHR+=Ǎ}j1j{"|uXXtyE >ڕH7!v~{ss۫_|L?M瘝xcbwpRDidp>R4n9j9z5/WjmpJ{\*ŴWXFx5$aⵄiE}8(Qڴk|“RjQ3 ?@A\2U5fp 2eUv) o?E3s9~2ߋnd RA BPQk 8 N&θU)04C,Y: 'U`6rĀʡS#W̽*:#8PځV 孂yDL.wi`$ϖo~6C 龪gmq6^wbzxBj_)MUuA*mf:'r )VY1c\د!}t:%c:G)98,ej"xHZb }DMї[Z={ek7m_kl*HDtRZK aDF% K 4&#Od ؊vokck HtEő0[WO00{tdd o&vt`Qš^EA:z,D?_dYp:WKewmRjvjCa6ݙ;w&6.cmIcnRi[)cnm2&f$pWLJA}'9,0˸j')%@jv S$ia$*F'9@ɬɉ'Z&4LpSch jLʚKjĥКGYO]G}g΂)JWB hbbRU}tbo}]q=Guk|1RCz0Nć \󙣇Q.F tMIVN)܏|8VE UyjRB|Fº1?D괼;*oǥ5X,QZ^jbv T 3r_|8xU-7m[YƓ/׋ku d[IFn-I%öbmǓqV =)s0,*oo,K]Ur[[ ͘_,Xy2KsMih4)}ޏ~Z>R2|`\^ ;7 |޼^_qXm~=EsTPުhnDn\[< TyK) ]OCُyoVxFͬY:y99Vl`W|KGYA!} GKJxV܁iosѣ5H4\%H ]ȴYR"\ ϩ`5Iϙ2_3*hMw6S/6(/.FϿڸrE2i.r\73%2<`RxpAH%3&ԙ3&tiNc2;;#M;ժm$n;JS} ]=&#mj}sYʸ4e1#ժDW\\!8y$*o)U:.V8'äQUn #x R&E3w}-u1p!+-ifI{x^o@G!/>GPpA\džQsۜYK՝YRԥDI*|S&t,tR鍦"K4w,E2u:ם[#o-~&#*\Njs:D*T4-lt=Ѡ5|eTOF_ `ec!.NLcSE;%;JѝUz+jr:ģm4+i+KYڲ%B9Pa9S7фg$ Q{ eJ,,)2bVŒOwy3)x,V5!Ǜ$pqn\a,2{+Q&3aNc``E.'"Z.(E]T*,+  ]!\&C+D+ZOWRΐӆDW<]!'8;]!ZNWҒΐҔ]!`1p--t( hmHte`@p:#\cC+@{tpBwtute6$c]!\B+D+I Q=%51\q } ˬfGs7Β^w $L)_ؚȉ}OAzY&;YfR %HEMCGDȃ*f; ^kfYCeܫ7 (.%RN(ԪȴaP`:OK!,= L9oz^LK&I3bIT(椷L8y%,uSM*$ىbOܕ {P/3^>kXN3u'lDwyY(mNby gUl]njX,eB=kjN(Gʜ|ia'zia7n1=>n(+˭XZ0;,-nia߮2]`CU0tp vC);:GbFN+~&MDBWѶ伣3+t2e< @e,BBb]!] K2$Bִm[’ K ]!\1hMJYMtH <B"vD[v(v'+m6 '+׮4]!]YBv8JWX` `A@kl;]!J.::G2Zb$OUJ`bv֬"8;{N:u5 =JIth@Ӷ}UOWX`  ]ZKD QR#inJWx` R ]!ZaNWmf]= ]q- T1pi05t(ig %] L L+á+1(&t(۶ՓЕF+]`srR ]!Z+D)s+Ř&!-`ct0tp-+DLGWgHW)%C+9 fњ;2 Jk;:C2pn+,e8tpU0~WV~ Q^Lfw'ke'sidȒX"q&g#ސD;Ew`pۭEKw5/nq20T~ Hɕ2,=H 5שr<)cG0nncw0j6v/u(uܧ0/$I+i4TzOghC@즃إ)po>G1)NxKwzpCJ)/QND>ְ?rŜ8rӚgKN^  REt$7vvkN鏓/%\מ1'(5OGukrjZL"`[uh@l6 *](E*ԋW>hX| /2Q9Th~̞דa6nQTF{@?$8D)v {4NV?ʕ(d_^E &Pͅ/M7M5}[o5wR~?SX%:q9˾ ku[جMnI,XQ\ꔹbY3eNiՏͩdC}Mr-7sҳzv\r-^} Ht՚њ=w%ęvkO8s7Dy~wDYkMhv4oSM  Մ%5]!]1-x v]!p Zv؎ΐ[ ;f'?ڽ\C+DZ]!J;:C'=z|(]!`k+ahW~ QʎΑT P rz Q i=]!nkGWϘzM =Qy`8,[FIR9ujycſ߿ _|/7C? vwdG{+Bo]Uvn):%I.JBF[q^AfnF8Io}Rԥŷ-%PQ1|悍N:ސ^b`]wx{n97/62_wT&rAW=U<_K0+(l}X2iᆹ={,[a;ٻn$WXyݹ2>h@Uݚٗek,NJ4IdI76ﰓTlRyt> />cn~{VG3g,+; >G2<i%6ӫmy?D'j]I[2z(ZBo[ds$oUn:[w݇}~Owuϟv7 ~FnWhq7iSnoH.j.%ozEZJځX嘜5!;(tRBGHWr=CMN*YUcf)ԩXRU+s5Z-^X5uVէ?07A:~[_j%n]UXsu&̷wɤZ0D3bKRkPTA*VZQ5m0k۫N=hSts6~-6-=5 ?\@RRΕܝZ1U]dlr))5zIEھ PC2c1^#1+>bnl.*KNIys/HDk3E2Oc߭ YWF@ۘR:WmΖٕ%cM":IFܱT`!?}4,4T{K윈=]QX~B#~9oB#fr'oj*uvNTC))IW6սs2Ju3:%D:f7EaNR -BB6oЎ[K8- V\j() j%$DH,8dk]Y+CԨ*͂E>iuRl U9]0dTW)NuOrE+dnJ9G),X XtfݠTGhOM\KͺnG0L#mTbˌBpvlp#)KQ[rݫ N7h WZa}S68mˋADUXu>:@2ŔgK'ǚȭl%ρE"%l0 ͍c+k.oeB`tlTf*(JƬG5C*D@ %8ջV(qMA'SX/H1^l<5]6Ajc4)*@F&0 β$ AQ)tJo%Cez|=rg3 d2emB AvE4rzg] R }Wh%WG e6 a.%HPED&T "=+FQC(d9A΂E΂G iB?%pB8 :)ԙ@sB:\uT&/ŏfUޛj(hS1)E@(͂EWP`ZtfI1;"JQвAw/uz62R6+QU.g-^TuA"V_R}w"r 9lABb~)Sl5J.!pBY$Y*k hPM41 =IB}f#aP. ŸKUU;f"cEU5!pBDEs#v d≩Ľ+0E^VtkA#^}s6=D2 o byPT8xiP˖Ld)H!JdAh2BCT .΃%-cy L91/t6@2&@vTm O=]/d!T?Q y E*ΩFۆnBZ!BTD/l{n9׏$qj`1ugllfѓEFظ:4%{hPIp¨]ôbw)DmzIhΓFꉡ057n( 3_ AzQ{20pл!)C^;$u5\Qv#"C[ŹCvѤD+9-TP=`ChuFCJX@2%fTi3 A)w-[F3 +EkMkׄ>Hg9B~&o(XV1^8P cF[TcD9Q:{$zuK[`rOuHMuc@sZv@[dnV47=֢U 6jP%hw&u"1KCbR{OkV/z&D3PU@TcUwUvAy V&-h-\kԋ+ črzGhi0kNrCɵa+V @_OE<` \t*΃el) ʝۥ!H0r(:fŠY 5L./BG!6SZ,tQՒ-* v޵'E<򸠵RT1|skovw},]qZaHSL-TnD '4^ `&L6!Pv}'/nxגf[ʇKh!lj[~ݾ\4Lg]vNn 2xqf{u6|>ͯ{8V^NoS<_StѩxzK|⧶0X/v;9~'uUU U3~tJ_`d[%٬ _IK9/ 1P\t(@E@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $NuzMNj@nX=;Jm tN8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@2U9\֫qUJc tN Ҟ8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@qhMN @!8(F@`+/N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8~s z6|b5o0Z>w)@gFMn7w7OTTqfl‹F/W;mЖ70Mnmi]?/YDž7FS^X3RSdGS $GhޅdE]qcW6Ob}k1D.$+si\I1DJ)ʓMk}M/,-Do-< VsOݷZx?zinki(ւyւ.y"+WCWk<] >]!]K 'WCW@ա@ҕšAv8tpi5@j N銢5VDW0o},IhNW当BWCW.u+v_a3h)pBW?x(]#]#i8̷p׳w5І+Ԋ2LzEt57CWkֻC ]]E5R"z2 _am!)eJ-tut5.ZҥFSgN)3 g-] mOo}ܯ';447{"iB]z7~EtjJk)ꃧ+$BWGHWF۠yEtekmE.wqּlnhd@ g,]gH=mQlkÙog<0c'p\ABW NW'(I/7`%CWW Jh>wJ(h#O(]1‘n7Lh:wJ(j'rsV9*xS**$m9h);Q4`B?J֪HW(gPSM3!)t򳧫R▮!])4kR=QW>*=g0$ztTrIOk]t8 -?{J(Ezt8fVB{T'y ޚ}U1MhS8zj YA?UF"ڽxQo+Ҿ>IDixQ/nDHڗ|@ ?ݡcىC '^#qZ_ŪB ǡdgZ`mhᾯ3Fo]%6\BW "NW +„At5FI=]%UWϑ(5LesUCW⳧!]1 +QsUW5ZΝJڪHWSX ˓~S/:-;{J(9kҕBi JqJp%o ]%Z;] .7Wrt|JrMn]%9t|TKCynut$t֪I H74F]ZĹUBIyKWϐHNBW .mL*}s NW %-]=Gxsb9@-^2{xU9VdqRT[2vƓP,艗ji+/ /2ȘA:$u>hNXV$zkxD(}B-6}DVWg|M2UM߽(݋(J+fpX<;&?OnܧcAj 1v L/ ZW-~,/'B>^ 7w)dfg)/"Rԟ߽d;"H)K@l=nv6%9oW/i>O Tk]x*&f޽ϻwBg`^qe>/! S%,s^?0|3 08^F2ѯE4 i };6,G~*Iau {7|(}Q=1đ濤7/(ͣ[i1xYuּQ1*.̡Ұޯ@厶[l6iteS+{))2 f3ܬoss5LaX6y|$b(.-EDUp;ox O= mjg01#,RFXp4U} 0^YH.nr0K~}-|Y2x6F:Yqy4w׷$ҍC_8:q!FKk?Il? \p<|4S՗柣OOlK RK tq-/1h,F|?lQBU^ep~~HnZeWt 7U5uLaYtOUyv-b-)yxt?ۺ HlO&]\7Leݡ:Zw^̰tG:cЧV" I}?.2 |f+\<)tH,u~3Ȩ?M^@/]g_af|30M7/uЭ -L_Y |aQSp IϘSn|*ocH="ҵKKSҕZi#uAafes-c@i$3Gj؆zYuԍo'o:er?qX FOi0ZC\ Gl )a kub*n*PN!ld<)VzSjʈhA #(H8HW*ù{wH"0e/n=tKސ{ӵ-/l}vzK Fd:Z]$[UV~gf>ic8vM,n4@]dA*pFJ]H˝=006ïi`Pű# =U M42%܃x-XhmB(3*h2:Q1,%! jSEL=68w:_?dzrO Uq2Ve癥ȆGX{ p/ZYS,OY#YgGyx!X!4R&?kAVafhj48HFA{l{TGҾoMy,~3ܖ$STIbahi|y6KbW{*~[Fv쵲2aDD&xLj#>bDÊM&{  8Dc_9aUUS.3?6ivۮVY'8VULhGȩ]p{&8rEC##>C1ap!syR(4yHty h'Tج4PN'J@G$wb: p p?Q(n.m;k8AJpU3#E2oD;gpOeZD-ǖ{Beya&ҚI?n& TR#YgxOErVFˆ0èHvapڐ4Si- Eⵊ( -J+C@+,[kJ ֔8f6ɛOO ufw?8KZ|:L1`vWz*V_; zi n*yNߌ˶zݶh+ίV; BU%)^YK*T#Pu1:qOЂ(> \f{*'0,.+YUje]ӂUy eIs0ob y:I3.F2`,37tW_M~݇߼cWw/a, C(6U?^pxєU4,*v@\YE]<5TGLܢQˑz1\~5#/CzxW#j+,Ȣ@MTm(>}WƧB(f P3;6oоWXn\Y gf4Cx^rÜQf YcI"!tX!_6->=hkVtae #c6Z0 !Y-6ZL1sҀΧsn,@B *Y9G,7nmmiT  vPί{B*8O>鄾X /CU|$5׫`EoHbzP1&n0U/qTe\2tHlYy I99\bր9[c4^sO*fM'w;oETq/9xwpe7˯YLrJأbEYf+tq`{clb[?'m4?.&.RgeCs;2]c)o"CHAlagyoN|w拯.休r=[?kj]ClXkZ/#V-Z 2eI;[Hmyiwd| [laHi"m)Ih푶NE}Smn4&no-oom14;4]Q4q,18͛-FELҨsCqŠ|?Y&,ߗɧAqΤM~W;xٻ6r$W|c 63i HxG<_[eLٝRwfV )d%jChqAiM;DIyd}C|l(D(s:# 0<0J`Z2lDs e <J5u&7tsLcCF E)V^nW=pkU=z m'@+|N||?"ut=*E<|LΛu=;ƧL>1$NYURe)[Prp,x=K"xtv=|wxА<'HI*F0B$mzԒFsA9/G y OQ}y4?73#7/caH*H9x#L cvVj DKB@k դ=VXdG š(CwuAJQm8e;YI_E}} :;a<#h"I%pD4Z%5I-.H?Ʒ ww/')~>$w~ۻ_I_~^f•w8Nt|ýr1~t~F^aJhONV}ǧohg5P@ibtv6Ou>Qrr0{zsE uG)ޯ'`ujo§p\$׿/Dt+ii'G~x:~p:>{ou_{ó0}m;k]{>m;Po8_,n/4q9Իaxqsxp%[vo6~5 Vwtv!Z8LH-Өy+#ݝ39*Bc}tG?9zC\#$ٲ/Q̈r~Lj FGY,!\kyWg$ƙtw~D>=z8 VEջ͕W$Isp+eo*鼡A/Ep7ӜGJfN{61fR\+_}67E9}}ctz#Wǽ_0.ot֗]L”\#xՕsUv p2;>_Ņ;"fassbmwdTL屿g^;u&k{D OI e2Ӱv{H>!/,J{&HR,9sLڒIĸg3 *y""*mU߼HFR(U喅̠fҤH hs4(:"J 3U2c>jzů^mnlů~5y+RQ/=V}ù ?i|~gnUW)2nd'1@5@ls[!0ȱd|ЉΠ !zGВ;ZؾeCo5b;Jު/ILNIkIB:#pǸ6z:F2pZ:C..rdƇ]]鼽?a`-Pef+YLkfR22cpEbDꅳvZעBrWKحv}&UuuuOM'c-V%rdcrd2lqsZ"<]q-s2=N9P|w%c_V&fyD)\ !œ#U6\%d3(E@#an[*Zh"TlبW=*C47i+OkW)t6CM >O]Y*?ҊU^FWMܢ w&L'4ovO ^<ŕ?Y"_kj[ жvдIӢUޏ ؓdS-KRgм.aAЍ F1o{ǻU_R@)D6@=eD4)`yCRVfcb^uw o5e L81KJJґ;\ZА-ܤ ClgV(8&` t$ T&%Fd'LQOHH9:( zTjݹ[PSKmH-M6OM25LM_׫TBUT{D8*a 8p#〖nh:ggaDOS29z dd"`1˒)M fDPbsoj\k#c܍tzfƶXh+cpDx+ ӳoHNN)wf/t<<=CaT3A)c29eǕV^B 9y-#͍@yFzKktJ+30c66QC,L. p 1dleĮ֝;L'fb͎mQUFm7* L$&_jb IȤr"CB#S#42qtɛx%"tL!M~i&}@+ DzNcz${$" ͢dQ(!MB1`hL.U#=pF]LP퍓p<"gS)i8@|B++պsM=I[gr''Lds #$&e]RtipF| MH 'I1?d NH '1nTVZ~0l Ce.If~W-) +M검rZUM敱!r}ܤ LXHȬ&;gܐX;_Zҥ~X ]eb.zR Vf$Y:q]s5}t37h[-x3hp>;0n{;C F/A3yz/ڿ^trɫoLCqL'}.a!;v%l]'7}\ogÒq4҅NI.ڞi&gQH vAҝ[$[8@B"Do$K:6_y5,{KJ:wIEBp |PW樻al`gGݶ l6w{@ ^}3@CO4{#2ڥ98xՓpy6Zh#^=Hvƒh%dbieJq&$뛠KJFr1p8<)k_>6 =jOV@e.Y22ql,uV.cF8i[f~8Y*Eks5w~?ʺP∋*ܟ*5"{WN]Wb}Ug{Zl/z t7~MkUd:'rc#c mXh,C4>cņBE-Cw@ :%CEc:X҈bͰx[YD=5"6sd+ + {+vo^Ⱦ!q41|6i#RcgmI 9D)E 0;ۨ\R\ɲE[P5yjWr6I#OmRe7lJB*$T% `qo\$*$R!I& S9=X\NW %f]BbRa{DW,UKH_*e끮^!]qF=UU_*R k+xR; ]%7t2uJ(9ҕT>iW WHvJzt4J0Ѯ\NBW T]+@Ijt5kZz(,xӈrޜ 9ެnTAeAr2GebQ70x8oxIx1(T(#-0^-xR U➕fm&`mlZS RGTx0bR ] P4>|ݣF GgƲ嘪M믚ڬ▱ 0 ~KlSa?U1\VPQV~f ^@ p7JlJ͑d,g*/s!Ƃ3)dD 1)3tny$zV7bJ`RGSJmΌ6>) Dc1b*=r->gNpEo^$/n\\ }w-=8Qt 4]= KR{ZBa1CBwj kKaig0N'˒('1*uVi 2:Zΐra!D!epJaj3jX$"﵌FMFSr1.WU(J׈n ̉p-r(uONj7.TDυZ( /aJ VHiP.Z&"'3E%sCR'V@.J 58H$*Ű[)=Z^ٽP܎fJBή>7XճɭYY;H)qVPj[^̯WSI֯2iL.*~ڎUKPJGQOYyw=bXŽ%fZbϤ n=0DĨGyxp#[#6Q(֢t$PG POp`q飶ěU`̙@Rm3ckpv3c{Jk\ؚdʅe.T>) Jc[\&xC1bU :n-&ɯf2)>MglR{"9@3s) 1DK` l1 Azm褊FӐ=@%$ڤB:Hd`!#a:0- 9I2cg7cbL̮vkq(kY[=h$8L)6Q *2(hdk+AI6x&FAX8SaBg ,d 1+ k|c(Q>rA2g7VF}qˆǁHP70t- fQk?>D=7N9$%Et$/E-jy-v ??je.`y^.nvqVp5MnL›r,OLOupO6](kϮs@nVt ]>3 m:\L{OrhCju^=LMnͪChC6wz^5yzأ絖a^[[^rj_Ro-Ϧ7e6k}5%/n|-oyN>ςOj7 /^ŀkijU蓝\n`1%̭` smȕIRdNCX~y%(#I5r|hH NG" HԔ1тtL+Cʘt~yjSy`/K MqXۮ>O@9x>vz{i_ j|VRC@!VXh \Ì;dQH0OsFxFH܈@ 11蕁.LQ3QZKMh $'f&L*9DĀ%t.LI"IJ8,NJ7֝b΂0liZv-m :ͮHJZ#y,.jE :DJn[Mx4\aշWaoXU kv_ !)C@=H)j۲!>lI_jn.4qg8d"뤃L]6R'@QRRZd ^`- F= ? o9B7?ςr o9zs-'Hq%΋EݛӣtyG<,_P8JW菣;wfJj Wnn1 nQG} g/)rfq䚐&~frY£~{:`y̎=zw= KG?}çޞ帘sY\YL? f&fn}fzUqܤb2vףM_饝,,߾}ʰnm5QoOn>'M 3 j4-4Nߞ^'"^o n~ ,eA~LK1?}YLNao=/If܄@ե?Hh.HV,jIJ\Q|5CWuZxm prdIuL3Ig@X;*:$YbS+f}O O˓j{*&ӇsuWj2(;~ieȒLuNv}~羘-<[ }yu_}_)7y%+|=nuziȫGPԏ4čw#u1(Vǰܱ(g4y./T,0G^Ę12sNĘ&tWptx7rMQO@6,My񏢘 +:ӅY,-痰p)awxb)r#xvsUxKWv~n>1&faʹ#3hi&S)sg^ϲNn5=hS1xb439A5<پ񬒙a[-uM*fHȈȅAt Y]$B udGx٭CrrPKZx e<pV }CQNz`]G1q[ FˁZ!$.eX\κrjkqhrqQꅝ:e(!e21X|n܊lb! 8[\.ozZȆF:fD.gAzMEd\0dC?Q64E }Ќ=QVAXeE8 @XhsL0c>P9Lgښ9_yhKfUDϬw}"/ko:& `x<dЂ-Ԃv[˗Z3Irw '7/]o]YtJp/D| t(b0BZD-x,j-cm(ΒSa.ufu͢U 9;e7]K8u3n랰α`n4^EJ)Ɩ'MKV^uwxNC3ے99說|lˊi- fUF.0IA|&WeMZd]κM&Df>go"iBU\"O56]c6)+tUdINuF!R(]ʕϳRr2McfkTyK><ĆRT)Pl >kk&I* _I[@AeP&[ aN8d̰"kASI $D(RRD1 u 3rAo|Y*L/DYc0Dw8sm#o(b dJ퍉̲ faD\>zwt#؄QAVJ.GZj,Pq׉WsV4W}!6mBl1ClEBjf6gܞsg>٭=> 9B32fp'fBPhHNuqh֡aDjSi /)*;=pX > -J6=!=fz0]fDRF!Q-(ElAo18jNb!i@~o՜6>Z?_s/oTYY h48ezSJisP#T9 Rd sS9IZfk|E|5DIDzfۄmdh,%I1.l"6J拓VdžW\u!X[4dʹ.v߿ guA<)]Z45VVAF VK5>ur |̺/vX!ԙQ$6HWR?_S=(]nSV8XG{H3CX 0:Hr"KUZY&~K:"f{iapC:6Ġ=vo AgJ,E&՞L*gMc!{R[}IEQdI ?g( ,>jMQ[.GEmU? 9!SH1X,t# K2ZЖ Nnx$d/I.J]HSo*ѥc[g<6}rmYjaBgeGfP 75\{qا/h603;vՈث*^mko|w벫2”l4hj<АH2\B-Kvdd"HSxtX(ȏ(%InzN 2E˵wZ>%C/4Q׊A6FQ)x 3D Yf1oJXJ }_g{ožtS[,t t);,R~Su,\nV1R(˿ei89RZJMdU( HL҈ct&1+f*!NCX*zSm2`[9-z-]/~Sw_|>~ŗWIzA2pU$)XJԬanR[o$|>ivKgG^ۛYkPطѸ"\dݝ]3i9Kn|d+FMA5`bm4Rƽ4q/jO<o dcr+bLmnYb$gH];6GI5ZNG5Ň5i Q UQ,5Vҵa9FP,-^Zq̈́w#"̧%wo}‰{ů%)a%8>2}jjLQ t0)*p0)*H}oLQ\r<4x3)+ΐvrygP٥fJ a(5'9"d JQ&`pMQB^Mo7fJrGk td%HyPt"pBΑ% iݴvǽb,:i3Uߏy?f?YBf߬Kb|VO22UXǽ3N7c}G6;(I FF6\|E6ny²[rgUeYI^R`/G"}V9E2AbAG'_W2"wOGh9N{hWZ/SooNː<^.,Ko6]NVTQiBJ3:*.uVRv, ޙ4y! )ct*NVR I;d/@rBѤRaIr۹;hp eڽScv-'!v0{0}_y$ .쨴n- ECں E M8@ ҺPZ7y+ZDi]g)RLRIR.YERF2 \ | 4h^Tb2: Q#ZCRM$ t !xlRV,V,ɉ;;#牘a)M/ӂ~s!˓ײ^k얕JWlb2G) Pf1V4 O&NJı*W:_hlSۋ&ڤɊY^F$u39 i9"{fl4V`?bqˏ U{Jq%T֨Yט)_(Qej$T1u{ Ų3'{i)gs (&xeq}Qj{/8RMO 5%%C`d=O:k^(|CQe-㓳 QemH!jL%rIIb$M$jF e1pu:sk&W'W>y7[{ROϫOǼǿ&ǭjA:̻ѯetlj bȜ&[SCAy|Sɗ\_ }z?/+?"'ϋ7B^_j$hUԆx35<5#]N:e=q#_3Fks?==^]kEٝ7~9_xj}C Kb.gr~vr:[[v92_z[.yhu$GhFcD0fڔǗS+XWjBO<809#eJlu1ɶQ[ZNS=/x}zyWN貌8Grt9{o/5'9(]v?O?֟_O~{?~8ɱfෑ`ݳIs 5f追Ckjho1&xh|׭0OSn[!NY\)i?~Mb(eLezQZbG Gi&bVzl2RT9X`"MEd)GKa#ظ۴qv칕3࿳?;~}s{iEo^apA3Rzf>w7W7E׿)xn-@|nbug_^j<9 Qe72oW ~BpZ-x(Wncxcчn'"Eo\vу/w5 Mg#twDu߿QH8d<՟Q?k6GLp>N`CLGnΙ9K<ϭ1Z鷑#/~+fυM=ǔMڢ q>dw,qWʼoea5be!,XYS/ ip5zWd٬WC W'+PrpsL.OXX2 W'+F(N!㮆fq5T W,؉p4rfP_*݆Sĕzl&3/-^--v &E=6#-SǹQB~eE0-4m~ГL!w[(O6 *puτ+Ր{f~Şrm6\}\pC :r=ς6qNW⬊WCj PPIqNT Byp5γ>볒T:pur.֮ 88; 0ƯAjq=>_K[4 \fjugp|N Ws W>LǾ-" Ri }/ z+wo*#J5g|wߗWW ]{BvII7 ~wA1wK]iEjD-ew f. n|IWWKI}hDyDhu:{rwW5bD孼k7\uϫGr`~S j`>OzrO=n_|g?cmk/#{l0>U ^?;qa>*@A; ݲg3pE 45 ^%~׻1GA\CAzk/œC?9Yu9d%P5)U(MnQq]H %)?^Viכ;IF|]Nu-VN7Cww_ W^$g{bLj8"79&e[=ldQhP5hsB1{BR*&UXSՖʶ3.$ݛRy_`(lRhtg?-4KgWjߺ16Rn"cMdTZ-hstݫVb cCB[%ε`0"4hFJ9Egε1, w`R瘟fw5DK]$rKF1=FI#Sm/IN11ZGfx̮E b{NHB%dލ=C".յT=cPvlhc2VDRC4  !ƬRN{'mvZ ߚ E ϐ^7F~n\~d|HlQFK{q9 9K,1`Ԝ^ 4fU!oZlM%U 3 $m9ӝz>C—z݌#)CDҎMѵHIK -jI!a##[$3nT>! !xKA"Cfj=Y T+AbR8>*C׌OrpSg_c!gEƥWuAgPH! S#׆R3u4!kG+/V :*RRb`W ۸=\w^ 4sUs;Va0p$py=/]] T8քꆣ+yXLĝe=( B fCSaa+3\4en@0)ƽbz$ƬC6l!C.D@ %ٻV$V*N!'[>x^@1HnAkҵpT+5%1RLl  sW CꛡD3w%e:*WXPclB2+8@>+lWd% r3J*q3Fͷ@tH #a,Lh ]uND52|ی5g84iA0(crN}/fĥ&.ɉFŘ"ySt"B!&D̿ 08a_wooX@a/Z@mDe|uj :@a{u ,,hLEQkYB3m>%T$cw 䁀:A:*b O 3jBPX-;Z=/p"cQ>f7c~6sPH$d ٕ`PA?A0j Q8TDVpZXT6ԆdP>f5Be1 ە%Vkр?tX+Ytg&QЀʬa+JTjN:z{9.whE^H"6J1N:YXr)DyŰaARPX;ʢ#B$KnF1t01tU=Xk~z,amL 18ƀd?{FRЗFOɼth`v34ږ2,9%YRe:= .^Y+|rK35zipfW^ޙ4N0 U#-LZUZ?!uY8f?S僢)ԇR zfjcPfZK]jUamp/V`B£>`u-8F,qha6gbZ@YY(z/$%nK$;Q2N:TXkơ(cbIPʺ*%A!n&X@.4D/}Ԙ-ꢬ4B;]I cWec)bVc6iTYUxHJd. X;*iAIK#k\y(е@oW4F!`BT 6Kh!/v-n9Kkem*%sZcX'o| 7F풳88zs4U(TH2ڀ0P}~7_ {Ck/uJ^`ͫz9$,ҮKZ)ٺH*~f:_}on~ЛnssIrwzZ_NNNrcI _iYzu*Œ\voYO?wwldڥn!,]Yes,c[*k0ZI)XC=]pW]J 1K rLgMj!z5t+-LZv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';_'8J 0u:N 9Œ@cc'Ѐ@?>@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N d7"%'QAt@(Ws&UZN!:2 N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@?&%'ÁhH rm:g< {(@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N8,9?\ZoRMwqfA?`uY50`I'e\B֧c\B^$c\B!#S"qiƥ>3:]<.6VP% ${Mڞd]B .dL$_)[WK1Smn?#c YGo(7nn^_mgbVO~N] zV1@]k[Y 1KOGM_g3ϫwMu/]&7FCZ>rBAo6Bnhϳh)[^?mzɛP;}zNqB팬2SM-<8y9В!x*_/Unܛ`Z.?/!s ygO5k%0Wחػn_a Yn/.><0L&x41hQ`HP`< \hjTP+k)ۖ3.LE.s(CQOs2ӺRԲ%!-{v* E^6RZ6uAP"+C*S ֚اHegRZPlz5BU/^qO.y&c᪟JL+Ÿzh+a k-5*\Zb q5@\iL W(X HTpEjϮHj)pR `* Hmq*HmzG v$+ɮP\p\JqeiVZe2D2SVqnQǸ" tpErm2"Ǝ+RXWWR HKg ҩBWjqvB2" ͹ϮH%&'fLTK I1~a섦m^g)O=vRO Ԇ#MREdc`݃Қ)ЦK+ VB&++U*"b4q5@\)- gXX"RuBƎ+Rjg '+k cy <"@{tBBJ+k\*"Ǝ+Tg\ WtMi0Ng0HrɮH7 UZ)WĕSz33x W$ɮH>"!ʁKkyU:\\T;;Hṫ!!! V$+kB*">vTc;q, F+ J$+YwEjW3de\ W^q*cvLGx2Mt->rbD!w=5N|G*gkYg-}VK/@um/'L0MH~ܱr+s$Q0o_>n_=mmѷ_ 1{ZPS .I֋5R畳dkMB".{}*BNqE*%υ W>!\` h ԱT*`\ W8RvEMHW$TpEj[?CĕwjH7 :̲SRϮHq4` %+d2HTv'yplN1d?c& ƠI{]3F{ی㏿ܶ` t?oD G23RbaL?wNʄpEOW( H Tc\ W*0*!\`s]>\Zcq5@\CpHi HT*!W'HW(X}K~rLWqE*c;q,B(H)BZdpErUHW֘qE*f\ WN`yl?ɵ I~Tø p҆p#|,\tpEjWR+j"ҦDO.TpEjmS2]YWσ`=TW[} dg0\p'EZy 0cկM4s-ڂ7@tDȠ՜3YyZYyiB'NwVDU(m_C⦱FFʦe*LtU7ʹli7cL벴5҅RY_&Lk>ÏN쓊ЛѯԆےӛÌQsOWb/?ūv.4ĸkゾ|E;{6tn~McyA &c~?AҸ)s(Cv Δ:ѝ6'w6_7g1>VZcoK=cVZ}%TZ׬1eyVMɲ *!\`mu2"j.?WW267!Ypd 6.\ܐ PU>v\J#WĕVZP32\\ zacTq5@\2&ȄpdpEr{}/. Ԍ!(IW(};^rg:G+RT;4q5@\YCqENk Hz;ڣ=jrV.!\`wm :\Z/B"ڇ+T)eW(؈tHLf0HjuS2 W)thtpEjMv3T͸"3\E>rbXhh^[gXHnJEUȍYEn~yMyk7XOǢmWS9= =\{ ]G?^jtT^{]Mo W[{o/U߮EuB@}URԭsuVot8\Tus^8_|x?,?hq͍O".o7\lo&xPN&F#lrjz}*LIo3~>;_gjlד;n 퇳ӟcP6eUU[+r=,yù R|X`'gV'wsi0B+5 tN5yFzo x8W=Eֽ/1QJ)ԽqԴjWVRm=gpYuP,yx{fBm0.fM Z*}{%zf:`;B;b@ RqqVʃf<9n77N~?mZg\sR2a ZvD zGmM'h;p: P0> H H-؇wۺ!tWX79;_X"8oWH&~ 3߱t\F|oo]pJmsEYT9>>ct>ߟ@iJUF (=q' ٕ>Yӧm=޽.h)كJwB ϰ?cO$WT'R 2TZ'0&-)z `ɵ:\mmp N|RӞ/5Ye>UA:@^)ܦ sP:{op֬OuPayrly7sx3-?s=ۇLy#aC%B^23hTY*]ieJ_0e۲TZ2"<4W"bZ?{Ƒ M~[YH6AlaIdHʉbR҈i/ݞ93_uU}]Sݕ3!t'Ͳ4ۆ`g4LҕaB֮DWyfQvGLk]`#É ZwB4&+L#򽑉`VP:ؖۺaEX{ݡdB)f׆L9L'͊ƾYP5ڠw FP5[t#ƥGr鄖z7(_;Pin\ZF@zTЀ 3b+h(tht&#Ji#]WҊ cP ;]!J>[]`H0tpy0wB!0t%9TCWP J;]!J#]B `YBBWV1 Qj骇t(DWS ]!\Ltho0zHWZ*CJ04B̄BWV Q]䖅 "`+ ONWƵ>ҕh]`+YB:+Dk_jR7HWmzZiMY5?(Y > rV;ȭG? "Ǯ7s{z37HIJԁimUO8f] ,튮5wB5htubJ+͜k+{whzHW\SY@t5;D]¥4BNWRHW=+!MwGWX "\cB+@%ՑzHW: .#wt(U\j#])$ -#\ ]ZC3(YLHW@.R;BCWWCWV[ QڸG2F k΃++y3hKRdte-!A]`FX0tpY0th RdtT/m@%P$Xni\r5L.v𰻇j[_٦6l}d(3Jqf5s]k9!o?) .\*b;Pʪr={f'mai >n(g01U=%Nk/nPL0x<|p φ䪞HfgۘXN;Y9oi"sdUJdku˪F'iY]9<N_oRftrw ScQtt|3gU OEY&8 ,'t jvI{d,y:ʧNx<lF`eȖ`A-(y7Dܷ4E[s>dB_~{P,Yv͝l d"'/^/95 Vx9;{Y}70GRfr֘6"!t6!B}6OiRrZi^3WW/ǻ[>܀ G^2EQ"G2n|83o}χa\'@-%;BAx]N/v1I~*;ӫ|:KC 0XN|\bFf׵./~'DFL$80iB M*(ܤ,yr/k[kj8߼EmOa'tg6I5 jflaWOnhߤ-o'i~X?I$t=N}覇H)`!LA?ޛ%%%e$7ȧWVey% YFwiWif4 y^R%-$]]rJs:XmVWucMeDAIJAcۼ~`l<ϫU[[2\/~_.b;4eؿ߼bZrjm:OᓮV[-Rިv𩼴~NSNQ*aI/%/7=/~Wo|3Mg'pqLU5d^=)SOxjysEbVA1J;o~O̮?KpaT y"2F*1Ĝ_?w;NLf۳SAO]6bVR+țKvmaGpnJ4i?h+S*L'jRi m2F d!/3jP4php6b@g{=5lڞgbC^7yc{jbq)WQ1 fӬ#VJ$0Fńi?׺Ly~Vɖȸ󼺽AjN=ϫe-4W&i@j΁$=&d 4ΐk)H\Q&.9) %g#Nh zь5oLLTK-6wYq58#Զ欢^+ ǭVaZnH0hA0q0nqGxެbT*߼*:νB06^=3`+7f2df_䀲eau[;%1[ory9ws L]:yM>@ϱ#OGyG8x<t~W-Lhif/pR)Zim0 wb$ Owyr{W_u÷& ݁m @pk>8Tbo?p0i-r@Ó'|"~մ 7bUqh^u+^lkn!V={1ڝʐ~ mCrLPo x]Ɓ:x.ݍsUkMn{G{g|Bq^;ߦ{킒}ޭn|@t"fnQNOhUiimd~4$1tsܤ!YÖJO}_H$/9gtfι!~vt3C-3Y׻6ٕ˜xp+E$v~fLqIuwϯ۹kxUX)?nm$5m`y^tUώjyT/}mvkr2}L! %d'L G@5!cu֎;c7|~ag$inw6~ɂPu6];R6 Ni}OԈH6N' +935,>lOW7%W7y=nWVFvcԿA͞S쒦vЇFEn`G+;$?_Gf+%kc[p\ŨyHic:;zQ=ݺͣSmکz=nʕxdQ5 "dыn4NV Eڲ5Y-_}%K$R*4\cթSgnMB8a`>wX\N3&I7 T%ė_ sskgpPڟ#)TWͦT O{Nڻ@Z#,1Ke#s &k?.Mҿ|[u9/Fů]?-5&P79-dWcMG *-dT0@%4Cc P<$Jq^Zfꕡ)NQ(3:"IEN@$LAV,0ۗnr,UuWnڃ*m28I$XC`fi 'o#TEŠ>J-8Ju?K,/K0_VϗfԩJȵY1P1 Tj%%;ɺ.*SvTj EfH P/"ʏ^4& JEDTp#gQ4+J08(IF'o Aq Pd2N-a3 JACsZ 3e:FY5̜qwmv=`mYxqA-Ld/C$nY0e~u:ҧ5=Lⱘƣl)FWB!v~q4'Zx˴;#텛/ JQ L8/Z TI  t =w׳t2+ԩ_:L-+oauhwAAqՑ+$osu^.hOc9cb[S*'B8# >GY/drRB;!2?$9v$2^~k$q$r%dx=f/Zi (iJ3 G' ;q`3`H9tAqT.vk\3c*Lc QpRʪg`5L8m!zd|P\ Rݖdt<\haqy0WFB=g ϸ6'IN sEc@}8B\zD@&~k=|j-ayc<}:|r3N_DaK@V%1<]3 yRXZ}SG]̮)C1x'/f07T9ʋ)s2(X2Q0yvx.@p(PD9xQٳLKb}a\8d\8[뼯]#N(՝+3z;nw- 1v/808B5m+Hu- {>8mY >[.8x 6qnm/snyg)x}x~v F`,IZc6Ky]7[}qsć85r2b݈9l&osH ] x0qXJE!2 sDK"u SfW~VfRP:br}lp3Gsڱ5Ke)WZQ4 Uq|Ao{l8D0uѬCsHȸ4L.*t:3| fcЌOA&t6(3FqA0r?nfP4p6Bs)|IU- 5r G_nRw}8joy2-ZL0 ɻ$$"%Ed%$I.t# y::Oڍ+e}qvo#Y0U!T5p>Ծ{HB 3ǝU!Ɂ$)4*Rm, Z8zI5fٳsMn*c?#|6߯PEH T :VȡhjS6m g0+0$C0s/ MO<f8(|Xj9J :S,I̯p5mCagx׬ϟYMKN6,T,/ͥ]%"td' PL&q^^URBR 2"jfIwK>јSn`>2X,{Va?^zNWq)eEvH ǰATwbZ=*#|g;tr:}zWitQ~Pl-]\'" Fy%9qJ꧂ ;;6g6e^ &AhUHU00khtWcF2j@ykVشnO`7_=9Z檇Psٌak5 `;D~Db,{DD>Zgԡ4dž[`~* _ߠADTUV*("O{FKNÞN w8DPaUeT&4PF6ukKqddzS%1.{|Ls kW˿|3dXƒ~6Я!J?Jz SZlzObUʕ^AȦ"/ὴpٯ1:};:èR2t{:0#paPwT Q0. F-.y tqXkB(q >B4YA 8]C9 ݍ>0x/ʇ?ny1^{BhihE6a|ð0"Ef/foh T/y~:+#b55jX0 ZF怒Ţ׹%;<y+绶(V_RJ5EjTϗmՅv,3p,ճE\a@$IhZ`_W`)xiŞM[yI[HlWIotgU⇝n̔4SyB]!%q v*G ?\PAqIUbUqU\Qk9b*JR^y>N;*ǹa;@a7's.x8i__Cw PkV8B)I_;)eF j.{!ًYK +Si5l&҉f,PDBBT(y`P%?R.B56̪ 7P(n#,{Yp$]LBQy8j1•J"`I?'.xCLk_@̤Oډ"!9E~GK xz^H'/0j幱i ^?pHk#0Y) gm$7\a`q۝ f/gX=R{+ْ%H<'@2dU,RSNK/dzHdQ8PE2䃧b!~ep s!3/(%N0 ̖^*K[<&N7W@ }L &Zƚ;o@t$K4Lgif281%eN4D娶Cue/{=XE1Ū۝X־D('ơN,shzܣ,'INxM/IKh(jqE8;͎V/M7!IODJO3ަ>XIަO ׯEO-\w _F]M8زwEJd`7[Fx 0d_慎uYݥAH?cH ,Q$Oanm>ɫmO|8Md8JR{(iR$jqX}S-XضCN8y--u*4A$M x\ K DEH[uOW&'.nm |9Fv%\E_43+ϩ(Xw1{s."1l^Ѕmio\}2G}Nkr|9'#_WέSbNA߸3JEmݹ`x#x:{(e4խ_'y oRv oU!/±4DHtg z;`+oglz|4WD1dt% #"t[v[OW-@s %4*?e˱zd":c P ?Ç.Ҝx_.u>v7`^u9aRoB |ٺu۳Pp>Eid0Fx9#Lqڧ26p82u`-Qqxǫ%qMû5p-0΂Un=/Kl+ZCQX2AYB^̈́yK ZF׵FؔՇ~p r1͛CZ#4?O=`Xp쫾kq;MR웍b͂#waNcnwZRh $1JuĜ㲹@΂3ɏ4'ޔ3ͩCKn- '$=1GΙڀc1 o[HZn(8%}6Tl *[-8яZy$LD8pc_ }s[}ŶZfca ZsI>$;o'I͢G FPڨ`f&#pW2uw ^ =|T $ 7#$j>r%58q\6+?=K wͽ׶]9RJ J# ,)ˤ|_- $ Z~XYb8PL%/A./:_~ԅv4ʟgL&WeZnݭ]PZ.•w ʢ)eM2Zy,{rѕKEL;gVEc '젳p".ng5ᚢ+5(cjXR?!:nV?3DT5H-5  X=Wt+s:'5!p괌RֆIq#b*d :Ẍ:ZyE:G9=E lŸE!ungЀyFʊ NDn }J)ULf)5턞3{>mV`GȩNZ)t~Ac&(c0܉Cַ@ڃ@GNf_?wӌF]#ty|ɳ4k$*8y;ۢ)<3D*֬~3e۠No{s&^8B?>L^CSׇx!Jܹeǧ0,j F!AgW'׭O)B (rt;CTjgբq$%)O|KMS(xSmTxs߀7;Kׅ&(۩J;m1Ɯ>>ۡ Tei ь ΅Tq61][[U 7D LR21OwqI8䡪F Eq2| d[_nqzuiḺ/S+ ~p\$+;,BW]AK L_aVIJr )IP" RԲGBi&x(b@;jqPj]b}Q2.VML=ު%Kqk2:*= վBW}*2^ =@evsx A;kV5f"hj8]Q֐qG]) %<4scOwK@;=(׹vALkpYR-Vr-<,@u_3pAVv/_<Գ8?.\.#׶ȵ2r]2X s%!I$1L832NDISƈy2$'@,M)p5ÌWzPg{Nw* <#@X\N~Va!1 B{F">KȒV DW#c" A%|z١d=2gfLqdEM`Ä' jI`up0E  b\{0O|>PGUv{reǗ %\^lQ2+t{)icsT5f,ܺaٖFʐߪ> ̅υ[Cqչ*qD;1: :5db`keqzsg$&߮ i?ö9/x<6/{.L;T< bm.f!#cF)0G? PnTb7b<ȋr.hmE`֜+"]x>S{KFݍklms{=b4V[pty[ctV Z=ELo4ɨ*wӌ N$ÈyVˬabn+YIJ1yW mDFK V.jQٰ Y;_wWNuVZ|'ᓟ`AhtY$9OZ6%7?dIhHL -cHA[P ޷(E٢^dzQׅe^Դv-/ 6i:ȅ-^47Bp4POIe1eȖ#[ve.-%ք)NӅg@8k8 =(Ė#rP ,Q$OԲGc{xSZ}EYՂ^lĺ{pN^">c``txnFt1wC<ӇQ&*56&e7 TO-z8Xm1)u8o&)-M m(ϋuhhe&yO?l~10᳔ֲEǃj0<.E<_73C)nM TC6{֦@3鿗b(]?7sp Տ?LyU?ɏ؃LEꏲSJG!u[R .: ^'8'' =cv{ĜI51PzPFjY((F֕l .j%גسi8>҇jZ!Cn(GB[ 0WAۚtù3){|o':@I;%] Ɣz#⳥Kz6,n+yg$Ϩ~)΄Qq-^4m(!MF ;ω5 #2oղ 2HhvbatQUV%If *Ah8)22-Z[`3[nQ1FԔrmekS.s@@vK@լ84@ r$1 [jIq3+fTF_h9lȫ-.'I7GSW?S ̀Mf\NшOo\0es:cщqMU\[+^:(\<}= b⹚$8V?V_Q>qi3"hSh];n&gB7.#00[+3ˢJAu& Ñ E8M(JSP)(TX^|ůEbpfRsn?N\6j&wtd8 Z8O:VgBacxJVD(GE%8U>SNWV ~h ?ΓRl̐TVctīeϙYlRwX:.n=0jM^LĶ֘ٻqWTZؗM<܇$UǤ I[mmԶ۩h3%(.)̃!g_:!tNfSF9r>ȇDIQ.p',u(R-QRix:NlN=!IF[iR8'^Ջ5FI_Ee(J(`],;{Vt^^RA+d] ht%%AD\{.6L\oy58ቕ`28 iм|ei}k{ئdǶV.U*N\ʧ}jqs 'haGER VF~FJM˛4)+Լ=>g%U\rChWM+Q}o_D cm"~r+W1Dv.`IcoEI)W7`ZR;1)u\̅<&w.IyiP2*5gͰBE5v!S_S8OU8q)Kc 4Ea HdZ{R"1)oNP ]b=k=N&O̼Y*@) 6H[.BD yw sc!A;sH ] 0EZlONIyBc n!dN¿<(h Kva ʋ >l JIWCV֋4( 2kvۢ\BI,R(D4mI)JlR|y9P#sG C(d*[l5lu7cs1 nх5sB"Lql0Yr =t2FT[4MJta0Fs5}΢v~jzN ?6НoFo!mQzy.⇆6w^Zk#@t;=uD0n{F쪵l|vrBx~-'mN*J) S"lX%#R$g te.D̹] (%?J@Bg>ԢU[_=T6p׻L:A*-2U֐T{sR?YH؀<>/٨H\ h˿M58co3W8>(mdlDK)bHlc['`uuۘ1td88I9 1Aý3SN(^zaM gFOE nsȹ|3tD J>.qpkF/?+VjJA:ϩNk@K Yp=4nOnYx/? J2\5Aq֤pU;8d`T=C4Bs#AEE[UAҧ*BfWT\@G\}fK}dFJ8Hd5%.a,|McS އ^B' UË˖˫ohjV)nQXVN&Ŏ\w"8Xc#,bV5]cCd˗K$=0+9[d:\ I |~Os[@3z-EDQ2T;n4j~tl悬 -qs@L~~>m9[ hܜEv-q` x& |oW+XT@48p576 66؏Rm6KK+%t]*|*N (T3qQSC{[[Mc|2^aR!QIÉLDIJ9HBj?rzelp{ӝ&\X|fKUCY^po"i8Aviަ3]&dy.4Le *.jMvxAq58fhzaٳv4>;Ʉ6$VA1eXcP۠|7dqҝ,M3Ļ$NP4,|4غYi@SePܤ x)"'R>~!&_랴LjsW*BBqvpB$!3b2ހ`z^*:-D\F\ne-Zx>x{U),ۍ2^oz#u63A}oG}2>ctl[)|W$zo^PXyu8 p rձ ^0sLXyGHm52rμ'3>@/e0[_4 'E׿w`X94YFfnF$+WhDx<}H8]xDljq`_p< E3[R ') l{v `݃ab*&( 2_L7 [o#|!+Cѽ`_=h_zf^{Ք{ef;Mʉ9>%%68HF_{gWWwM .8{K2l_{wp,` ƽc7&iou }E;uC%n7p'ӟCaRBd:t6&(/mo~M>Haaj.sOeb` MV0&\S'0%1p 08f~r`oڔL@0hw&r<_LZRϹߖvZty7[TZji V@b˯>9Bo@~1I&E-5ƮuH㯽`{s{UOqx4WaxjSs25p74cT `+8!Jcyx&dZ4 EӀ34k}rSǮn's[8=iԀu%[B˧ɷ|gm ARsZM'qeOL)LP;n:EP%p\ 5%eَ @g[GWzNi 9>!>]R&H< lISNZeS χCaX|➒&:pgMAP6T6oPB3TM4mf# ݥ;䌹{Aon&q0L5Y6uAQ6u>6ѐ8n33;w1JTrbU֪;v0?,yq`LFH1l,l62d[b9h0rێB<*у.=CHW /ːkjXȹ/T(q{#!<)$K6SR8 (hњsMer$i ƌN2֌v+n&5 CHoWCG5κۦF=G%Ñp A!xIC) "naM g;5z̴ Tdf$&E-Tϒ]❯#BîH2P)`˿؉|3'K@ mFYJf۝",١SOB1+iauVEnBӢLQg%OW4nN6U<Ask,(8d78΃ntn=-ћJL!FczR4["Ք FBȦ:tk߀rF8/dsm( IWCK%2o% X@=V':m/btd9$jfumE:^Ƞz!k,&t!Z7RE{gj1PmP$nB4:Rȃung+8Ec``:0Ao-:>N{͓gV kJd'u3w̰6ym.ݼ&J5;d%&wBْÕ hbUE6“BTQLm+i7rKBx٪ҳՂ fuR"ʗc: 1?AxCKv}xy#%uK74̞.t-ܱkbnGשfS9RVVd"iu!WUyt<>4?#XK"~xgBXF_Q렠ijRoQX-ih^";JP 2ysaű~iMc;]`/?2Xcjz~j4(L>f; Iڴ\_)`r{ rðiKOnk#X^{}>)wU_*,*fֽڪi.hsqvv}WjrَTf081Jϻ? @fqe!iLForh>Jès砟O݇1-1 9$QF&3_R(tFrHsm $/Ӆ?(F9;TIxrPAN(ۭa sXgPmcRFqÔhuiUWf0:9'@V]s ~ cճQՌ½ eL֚;[9Lް^sѯegxM?G?{WIno{0 ~X*#RCŪ̔RZUHFq|CeS'p% hfdU󕚻uT6`ilUx\f]-cjr{dҷw=f a!;T@#fkKnr/lNI PMjXN{P[OVO 5K@mc P@N,akmH/جLN?5Oү3LmV W//4ė6]۬aD'%Sз?GCG:ii!K?ŪTyOsg9woKq{URZw&-+4>0뭝&Oڒ:zk-rq~hGPޚloZ%̹(gV-*?~kWo> o<3ﺿ&^U[+h|[.Rh,^&%Z; `Q\=xvO#ᨤ>d&F.ˏׄ&dI hꯤi?VQ^(1ӛ|<9n[֬ѥ7̾i"''QD^ma^D fkhC;OS*,q ܃?x`ɋ/ڱlo9k C”^$%Όj>L.N׷bSы9>qA߰<|u~s Dv_*L{Voq<.kWOH5snZ?tݠm3m a`50oņX-$;!\8w<=ǭM]ԂzH)5hӅ`Q0UlfFE}ëx-oֽE+~WאnJRҮ\ -Ea_ /Huߊl(0U70ziy5 B =KPDa j>غ mJVteFU}?r.:Zp,9>[Aġ CRy򥝖C?o7lHu]X庡!bCp(I^OH-ŧ7^LFle5[F̆wnPlngsY|Uċhte˛H{ VИ >9=;9wqKɬAު}O =6:?I{l/E]2n`.UyOP}T60KBljƬ8J$}&wF_2]=O 78Rm!r]oҺܿR9sr8!k %VJK jd 1p檯D.Dw>]LRV\ŚIaltP57َ䮾p}ViYcJW3Niܨ,5ekxcK=-8]\[yd!zyvpJZnZq Y++uۇۣzQUUB1YY݄J+ܕre#(}Oy|~^yFͣˀ+VR1eTp_F_R}2ճ0dqs$IO YO3: kY]G_ 1dgjO?[՜ȣ\ y| OS3mGZt GqQ)@mz<:  SqX"Y]dc) #;gOi4M&zeew[ <$ z/!OZKu{Aii=xzg&F^&8#pJbx=1<ܟʂ`Mg6r=[E=R[I֝sb)UaѿaLP/䴦2eW(FKk%HQpLF,taI.E;D; B~^\MXh~# k`M} AKcY j?eJ%JY Uo\ T;!5z=F ށƕECbȥ;K_IYX$+h/F7a/>toPeYϼ2);_xPynVJ>$u*8';ʮb:h=;:ja3x p;Ъ p$HTYTZ31LIeW|G' 94YIA#n6TZyq^pA$ǀл`D_8A$vQV2|PIdV ak؍G;#]N;iB 3 Do;5][7Dv9W*kQaֽhۉ 1o3!mi# *@ ϋFo\q50 1:GKR\rR,,y!goFr<& 6_+FhDkXP EQˑeVpLQ 1 މD*A B>SrBso!)aVRдi^.E&3 QnzŦ9eU[+7<{tDEi'HɀiΙXR[,֠\ػ k Nx 6KMxG0OF+> x1V;8ﴢ#aGPƥc($ ֬Qjz@W4j<KeBtzQ,!2%$J;A[u+ ,c82Y WkhM``)u$sj+>dvqAy=<クGMF+ǬUTADG葙LJ3)Ϥ\eZf b17H0fN@)%QIeP)Goz4GW+=``ȩ֣DK]Ub{ڲRKJuaswݟb)cs([ Ef~-P\Ro^}+>-MLe/E+ *6A~/}KJl5Ad\(Ƃ*5-,oVzް%%SUoH\\(. C OJBc{ӕ;-7NLI.bgM{N,&qdg8,.`ŧYMR^[CERB%l0W$c 1a$%yЯYLҘ?b{lָ"g) [[ħzO#yStJm@z' 9 %&2R;R*rVDŽY91`TϘBUV𐛛k8ӕB?Ss;&rv6id*KdaA9zptSOq_zkPݳvLImf%^9z.X}Bg·$m_ߡ%KG#9rnbq8|!gfM97y愙GzqG+f Z+Wbd5i U! W"ڠ(%5ϸEpN?,l;NNOn"Ey[+O2TZA56 ̧z((_q$ ]o0wLխj#Y y!NtUp);kq>mywOsf¯׻-BR1On n@_'j'#@-⯟ڤj2w1Tݻ5}W Cl?Zi6DvTgy p8?5}Cfwpu!h2{X澋D*^ˠzT]a7t2<,Ѵ#^Gq2b6c0ZayAlnyϰE}EY;ciC^Lw'][L.">P #AEfuW^@MY3;>2sA~׺t+e{@Q 蕒+0Yh2?H) xA- 0_\^Vϧ?w0zԛ˟o;ߟ߿\l3Ȏh'Ux#Ƒ{7ޡ~Qdvo8QfU䤽#uo{c@-lO.+}ղRE|:^c9hP?_VMo>(i+C38+X*GD]u"^i(кғs#{ytl E>fgh$h S@ eؑY 'dJA1=;˼uJbmXk!X$U;dշjgBoNRRq"MkJq4NR{҂"pRKp3gRaT<{ '%*Kmft*֔NNT PA]օ8 iegL}Ф̓vmoXy A3 *P|؀V_;enhשm'C*(yj)MڴMZGks`Cٔf3tlflCP*ӶvDhUC9GXZVh0)5 *YIoS4I;>~?dSddU64E@q]o5u(AV" Tlۂ]9@0 mgKjSyuS{LԎeX+EĽ^XvA 2'~MOJu{ lg8yMvl[a5 [҃n=Z=:_>:KS,`KUf1Dy آU^ %r"ڦ k%R*d(5Žbr9`<:AhShA{It8MҋT{vmd$GHY$ޓd>M~ȠCW1I>}_tjmti~̀WaT=:72d'TWyઁ:]Mm4@-\jS贱~l+kBѴ;AܳҢzh#ꇎ?AdWOXGvVz'O<}LΙ*>B:Z9#R~u{N",U,cw',HsgvL,ȳk)G0O=}, _,~PԾ]xj5TZ -jXLڤ䤕  zUWԻMw+uvUVƛ]6Ė}C=ח"@54d頇DF\R!5R}h+X iM\ TGk0*r68)7hVI++UźYE]N]T\,W,윕546k®(w_DYRb IU1lR 1"HCUa1L+aT /ԛE!To(l,:7 $gvjFg?%2?hN:`4yJiU0o$@ #r(i!R<`K B6P3UUײ& .rr% _kZkZa\&ZlIC/y(2M6"f0ab oBo<ͽ; +qb0y_o~!+3KxCO7S0g[\!+~vGD7fNm^;sLʈj̓N:{5t$R>|\DpwRMJV# @T% WiU/\(au%ҭkK^$1J-'*xmw\W8I = ]ȌS˹m|y=?=znn NbcYmʡaWRS0&(K3~>|ZOxf?QIO&$@j;Vǚ{y= 4}0E**́˻}PN9i樨`QgySNC;ɨLn+%RўW +'D;[o{svVGtBa1:;yF1jGufXM7uwnH߽J+T<> k Y#0ܪ` <;7k3OJcbZF.Y1l]z>jE]vBd:`6jg#ԮSq|iϝ4>"Ȉn4F04)CujywU=zRK3N%]a֓ ez^#M= >-FD̈5_& (dCz'Nk}o' 56lҰd#!1@yH2~X(V89$,f.C ݜ\?ri'Ϸ%2뚻jݾQ/0>6W|?󹰟xZ־1ȏ5T}5R S4޵Ƒ#"X#v6`; O[mVImw7X*IRUY_ ,Xd0œ':%EYZz@8K7α&ܙBE'zFN.gJ>g_˝9w| ]0EG?_˫Qh+.ߖ\s)^iC`ܑ7T DN8$,B>&h2ENڰbmWKIwNT],1EӸc'um/|d~J]UĠ;ں2T9-go{HHV@󧷿OZTj5 Q" M]~c2]c7ݛnTU-o N-mTNu)gMu 3{hlpB d#7lܽ`7ese ?dYHuP̍AgcYyg/YSGp`q[W $==L&97ߘ\zkbX]rɞivB[@^rwxЂZg5BTi V>_ 5mItD%DP OՔC\[ $0ˆVAGFjk@vlsl-Rn`u4A7#֜Hr.n狂QvX؊D叺Ld2<`N.nV(9歕wHZjN9j'QAs*cNMlG9!0p{)ޡ z91D_p<}aVd%A7Cdhz3{0Xּ32%3G#@c2d<vY}Ju1UR̬D^z)J3V"5rqbٞn"DuJ$|Z86 D" jKP x$F+8b3Z[B) 06d9]=Lv,M}Ѥԗ}Ĭ L'p?UȤPmJh,Ƥ"Tlyԁ^Rfg߲#+pFj<AVDOYʝY+F,*N{b5uF d (jZ фfjRXbU0)PFO`W+Ug}dQ/97NM-./KfO}RlAr(΢sг}h o墶pܵUc??_O"4 ܒXߗKVa+~gx0Q" xE~TWgy\3a?N8/, @J>zՓmsGYe!RcX\x_ڎUG1u,ρbaL#enqLG=q˞l=Uoc! VqX4U7+V#3'GK,Ľ:2BV 'O>{3)%T,nD1[rQˏM#" C6-ڢTBp^@ 󲓄fo4HX])gߪq@Z+1DcsG-Wba@}?zޔN 5hZ,NPqlň$2s_0eNJHC-Ad37a!ؘl_5n#Phx:o:uܒvJ|8׋qәmCʠBSJcUa66D!+M^()ɏl 5X٢lJ KSĚieJjkhCʌHVh:#p@=vVx3薎CϤ'h#1(8DxxuŲwd` j8APAբvD)Nul(hS< Piz,d:5WE-1'!ocýŇ|C, Eb7H_?ȯd h>%jFjK@'M%I$Dj d5KY0E0&ѐ+n0Ѧfc{_h.k /n/kpsC<[id ŞKVlTFqsbItd,ҽ/9,N7ܯW;6މmb'`E$k֟ VUUggg~ӷ}MM^RP&ELQr'"(Oج3X/%F-$F5ܠ" UT -(S}cX@\l Mc4jsf iFzLa[ϓxc MvkeP5 Pc(+XLVa.M) \TӛcC6F95]yR&0K܄ƈ. KVɦdVsJ`FL `ЊR2Js*RV*Ht5!A/&&1-E}6|]0-X'Н9f.[DH\` o{ŰpGȇDS'k ?f^Z-}1uV?)tYI`j"CLƑ{3`[^I8vqi KOeou AzW-=Xv`Hz%m-.=o 9L"p֮yd1a,sV( ݨ`AN;oڭQv;[#3Cw&gp>DaYn|?ԭoq薐D\ v'D_A1!dQ4橵c9Jh`nB`P_]GƁgz=vGA=-5@==|,P廗N6jv05C- ~WK}lqj; 0lL7g;+[B>|&YzxXaV<h\=A9<]![s— ~!PŢx(漃֎曤f{aX{WfhmB;F0?3{ aHy΋ڞ >"fh@;َE~vK1g*hڡW Pgh7$:yh'vB ;r1fd7#풏Kfdecsb.սCfd7(߄sqd'^H=̽ې3{pqG=;CB*!_JsW5;#>3둃i-/\fy=ZW#V Lztip܏fQ_RO0{_TZ}X!.DvGge1Z[oUCGB(w;mMGjZOyc)%P`^9A^qbd>*P!aIbn LUVuJmld-Iݻ['N05CZ. ECP0Շ&Ɍ}z8JCh]e=ra,BseOR6-'|#rGre-d9251QC񟚋f37 ;(>!(͏:}m)-RX{;9EfV qJ+'s lm]8*l1vߥ;͖EbsEKc'h .>5FQ98sit2 9rs%2vn":'[iŻ/shWx:wscp,UoXŻunG#u3Y^6fΛT"aQ0~h][fHB/ڑ)A;ooKW ׭WhwcU\yv/cGۋ+ۘ}3v7k7aҡu#s|Jl#ks6h. VhwYYw[THQgfhMڭ.xfv_&'>Sv0kN VhwS73v(kE3&f1I WhByưB/ک=>X Q)SooM?hMef}\h4a:dncл kaB*fVh8^$-Ee ld'C v]d^ܶݝ3= QV-=|,&w8dn/'p9\0e,Zje}3a}2+̱IUuo#)>s4[aaCw( ]̏J:BL0>6YYn EFWʔ_tTg*\*xCekɎVoQY`$z"Ycꓤ4e=O_/wvJH-TGkP[qԇS 42.UkzBTh̰(F9yxJֽOnv5]/]5Z89o' !qwĥwyxw@8\?Э{/ `o yc"{TSs\33 猭bΉ D1`ݵέKh$-&8pYAi/|8uOg$U)<(JJ 1aʙKj2ıUNHA$l'iq8g5?hx޹CziAn|>F3~|3 ~ʛu㭮ZsOK6V7lMÏnm8j^m1ܙ6/kk! >C2/\\^qvf6o6q͟9KՋ_߮d?#*=rP` v꣸73׼~g 0ߤZ%w+pף\h:4)D6WkA] KL3YX%+h<7սRI8yn}G S? JGêzeI70ަ ヨ%C8AfJ#([y7{eCo|{H}8]`؃O`uJؐohl9͜5iUre=i tRCWҁSS?`+*}w[Έ3g}|W+?_lkba.";O!L!2n2Ra\Ah[Q7t:>zjkH@3msS,&tW2[?ZIB8%;YU9dnəNVgt}] ==,ᗝ1c gI#y k5~ \.(yʈ Ak]MC\ѿEt N5>EtF)<3y ^5{,_Dlo>#x{mqyϚ$rbaߐ bN^Wh~xLdx 例q13In|=25OnCbYci KǬ)'2ވtf+@1]ّeb:FLt&)(d3~:\{ UЅ 6j 3#kdwNKA~;Fv  }M\ۈ$ը0aR͞fkDlVFd1K͕%7XucV3Xݒ?f'~0 yjp:jNÄ*o텦S{C )IȆKN nN6Ն7 R 06U=wy.&nԌ|LT=#ٷp䫁=F3e>qvvqc]m<9Md3jZׂS-H/P-o;b EqMظ)y9Л7e@F~If%C4aT~~EYh;(#F%Km;Zh Fʌ%5$lLpc"-w7/볱yf_VcGj,no՚`oDz&͒j(6F\TRo\#D UjnKntX۲"{$V^#j::̄E;mNկn]2ښm1QL Hh/l]`%wOaZvk `\߱㬐}q vKtoC6[{oj-mۻ^wG]54;L;ڎ7spnq޵q+#;\q _τ 2әS bhyAȩ s,UۉAldp jw)ŖSSI/pl& uP)R]J+RƏyٻF$+B,RgfDFd1<س00uU˭Eݳ(uQ,JERlͰvK,VUFd8ѷU@J+SA (MX]>%m|櫉Q"pʇ%]겶}j+n}UQE2OoX|lNz$:x/*TꁦBq=UWVMj7N&nSٖ:*H O¶j뚮h] %2)Su#OW bw(pڔsЇʸ30U嫺|7IDZ4a WN_rQD+Aꆁ2eS!5 υ<^LSJqk5^7=rS $ic;ϋN. _t$5-U EU*\-&w0jlACV%umЯ7QZjR'*乪9W96|S%;>[ 416]mN.FQyH~9}̼wXXvVtc4zέ͸xOŪaqXI^^mD,-RT4bybu$]q}D٢p7{~/Q_KF':x* xDAm}a?"(5Sv50#rl&W7duǣ{Q: r{TyxʌU< nP/mv̾n?ήut?_/"o}DMXN->x ?b(ojLoէ_m8?l(9s߲$Ӄ.2+:XbGE U(D*6ֻfX4#{ڶnb"oA!5[C}zG[-=zHݡɰG0jC?9:vGMeM Dv{N/p.4y^8whE=}YpDz:Q _' _nQİ!06CyK՝A]s> g/wlnW]'&?sp*~2"7ʻѨݞS;B4}3w8f@uѨQЗ Y SwޚQ) y\Q;vhNCvA[S4y(3jh{[j0ץ2j7sy ̱\7&FvFM0)v{NԍɻCv)Ws˥)3j E BsK't1 CPU)Au஬ \ַ"ElVQ~GmcN:ؐzf,ꄿ= @:lCSK~ˠ sIn\rWFRy(s'QytMЛIa:MS& {XƉ_nU髺P+PFJC]ֻ\5TE1TN71,cAIj9̻4Gյm"餬vI1Zϧ  6UVIцuh8k{|'ڴ5u̶(3"NqU]g q~-LUw+ٛÝo/dKN6uOJu::r+v2r /lXCd[(f{" Fa/:⋰"l$Z?&bEhDte32sp}>^ݰOʞ0pHX}~5=1* zV1c'vsx#j9-v TZȀ-n}15^^\_5%O|d+> K`+G;.5{#5hblLgܙ5hOPuđw*k0!UVA(>( $4yY\hKBAHŨ<% Xd$o F =s 7VioX Y|*/YWwΊ1KY%y `><#nY9++U*; {b^9yAYAOSި~R;u#9jǚ] #hΨ3L?vWQ'0kN2LC JQ%ly8pH.hѧx6{:puPqvg[Q?પ(RMq>rzni]ޟJ5DDU2 Do G-YؑdA`RNP8 qC| *Z0oʸVT<hVaLLZt6(qrڴUEϲG:I,i&44w&Oy J5_znKMEPXԎ|gw^MU]/Yq}O4tԺj>,U+=Xrl3 nS١&ds\ߝ],tjZ(LR9VQ%i$i5 N#t3ioUAD\.JĮb3 ݠeM5u@W>[耮䤼>8CoSzo{a\ @ZKW/[c4bpwMkN/Սg'nyK^<=2]{1L.k1ǮIQJ.P*nSѿ(\&Wm1sMBjR63ig㟾RHtKΛc'}r:=ɇI?Ai럶zx eo'|>JUhӆ5pR(c+X,y0;\m#!!by1\(.yݣ>{y ?no~}4S ފ8RJxǠU;^^-~MĎMĻ# fXX}goF6(YE&\E7|HJ:l>y)r٬*-!1s=m:@vIyTW_#S;*swo԰m'o̫'RDbG ujZb1rCb-tCB.l;z,Xoy'%XpXt {<j5Yb Y{Wh\R@a.HQ;v "lnϩ2!Q;d_{%vFC_ݮP;uxکUSD0sQ (he8-7 UOjndrC.<=@3j CFvک;XI'09 E7 [5j߈ZS;j\rH;ac/ɨQͮP;ΛQy 2XNķ$vMԍ1!Q;żiRHNs]vFC_Oڽ-KJmS @ I\h糍 㛄h NԍQ!Q;x2yhF->ѰBÕFƿ?,$+=+4e~ z %:f;viWOՍ?ݍ:o~PK҇!k,}k,nPߜ] $6/uM(;2AUK[]WH5[b8#!';"ך G>hZC=z5t 4GiuyogNׇNk15?|\;+Yg [e)VA:[r/R,K7H{nD8m-Χ~`OT,KyIOu͊Я!E4_'z:6hWJti26@f4=9_jRZi)IVs{)u"U􄒢ZWб bNCVVj( + j\Cl޶cvnqm#܍|v\X˵l\ ~=L]{uNX= TRGmul5]bck;kv b>X ~!%u:[_Ɍ)bqۗ8ށTP Sz}!15yb }vԷcM-/S Sy|BT!!9ɓ)m=U "|nΧA XBb7:|B![*j`n(e0 0'NH79S7R:(jG(p,ިQZg SwD;8vQ]=r3f7 oDvTkߙQGx OR#3f AJ3$Y1'0h a.`JiCvFM[jnߩQ!Q;<$F0/`KF힁E&+Nب ?~@F[bnϩ11SKe((3%v@_@~?O?;?wûl:'?7 h,цiT[uwNGt0;}M8򲩮<UR@GMAS>* `5 4Ey[0A JʝKyKMX,}w-򸾹# ?: nG٥\o1)iorGy|k~?u7*J@m; jSco;,:lKT 6V&ʘ^| ;flv^0ۛ_Yz*'i?9̺^ќݼl"$Hܶ2t J5;?{0x̝vW8}[e%^򌎁x?{OƑ_!evgR݇X8a;3nN1E)$#A&%uHEQwӏ:)߶` VʼnT,[C39'pH&i3okDžΊ]TTzS? iVSsӏ؅B {T}S qv6@&fvzOrZH-,' P-- _@ӼL +E{wy[+]Օ2#RoI5$kv؟|-*zȷMF:DrYf@dOՖ=Z}_ ? @(j{GZBQq [}%ߡD. H= \AMH=h.L~oAB&AZ0!)}DbWM\(ٞGzmNuLK5";Ю'S9 t=x%=)x D $RU8ͥ^#1_ p= TcŚwhV7=+ǑD.NG.g&fp]31BfZz7B"H6-ܣ$˫e }e})`Ё#`JNYSwӊ3vuI? M3|YmNk@iT哴CH%QJ19f7ItOј/E r?Iܞ6jWM'3yz OY'z##';"vm@ ɶ=ᾁ[)i\?8~ ]2*0xJ|}S* !諓 xG N|ĤmV =t.(W?E'|EZ (|=i> AI|Yy{Q``1D1 bz?{^r-" `BK 1 [G! essr)'jaR|> {YقtJ Hk#& R"R:#Y-׆9y6DO1Ze,M SOaWnDWQ̾z `l5N50҉DH$XQ.mЂ)X#K噏*[(I+DDj x5U>bP-c1[S"`A#]Xn5h1Nc A5DxpRe5֖k=c`| ǐw.@ȱA# wYQJ+o 3E $/Bz1(`x7Ay#>ʕSzȿfzSܟ\NvYluU?/R zX>2 |s/͸%'q T:10c/L.t5Y|Pkhj8%Jې.PĒBk0X#mX=&Æܳ9tnn0G1BSq| ~LW5` q:9&5Wo8y9k0=84Pl^TGV'Vk ɳ6`jENwο"PGPH֧v[OAh"҈PkqZ0j*?#Vwc:URq1qL + ~~;20tߕfa[]lV-<^'r\E.;`^ oIඌ'UTQ ?ZNٽ|2O<^;1aUpeXJ Z溟bji{I_t:C9X`Q Xt%fLSB^] j3Ԛ6v;wg`1O* 6!v!>#ZraK0) DEtdA S/]jK4H yqŭ4RZCD3)dce,e͹YS%5t竌 {/YG=MbcR=OQNZ9FZCKK?xCRqv(igL4?}ZY$EjG|Z8tsNE;hɑs$VOssÖX -4S=rSxeRjdK\lt58HCZ>tw#4C;HJkFnuIEFм竏=>sTsLXh$3"zzȰu8Y4\ u2ych{! iA0)z{ӹ\O@M/tޏ}li^Χ' ;ʆi t03%%%^iG2yEfĻ"aƥ$h : 64}QVA7fH4!m2WMYS[VNS[N1F7k4?IZD.%mxt/ב["A`Ec @**vN1D-o@[8# p ̡6N (oKȠ{s8& â0)Ga1ҀEH #Vs`QS.(эNZ`dwi 5EY1ܗoej4ۍNƁ}$MP.#V79cXx,yzk˗G¯rg.27PF /Ǔ<%kq(& ~];#p fzΓhrAp 3PNOOk)Sr2VLQxn7%U{'"[uVi&(v)n7ݙ| ʺ%cM7h0!1^"ld<Ų3HrT`l(K3Q;&4jivws۫No. Tk-*(,k~JwHOHT]/w<^`6G_t=YJ IYkVN2@PZ[qbMeP~Lm M!J>{WnP8\ۤիwWۧӾF.~(7CbtG m襅j*vJVp LdBHôؐb~qG?Z5})HDQp4l#Ҭ8jIƞW}%t ּPKƫG 1l5l+Sj wgf%!(ޱ X4/ː?:| ߌ'gn4Z>+]R`C\ba\ q EGx|wMI7يSp[>dV 7𜞢&=zeɠ"өLFծA<3ҔQ-q]n;(l1sٴB,P\!ozE7{@G nZߗpx3.c93:+ߨ>aa,䨝!F(ͼ!Иu/sbGX(M3 baN{>!yV.&|tjeiߛn^?{sN3vd:-#B<-@y pY04Pf9A)v@E[PZ^a~;[@35RSwyσU,: +X-Z .pPDbQZȱɓWؤy1͐IL4QˮF[q6'sV{c@0 FkxٻrW?8<F^`<ܹ=^vOZ죝gd[@YUd-X@m"?ae 0١_?]m=R3\-\ ߙT/HJO724-]e=A2L`v2D%F ]dA4cT(*:N̿߮hppSih@-2ǿNprjUhL'òo i7ӝ!~!W8OW_|㷫E~0G9,v)z6딍zH/I2+"M^팇%2g>mC@rיY@[貙K)[Jvμ!*> !|n>bKܗn?B %tyQsH$hZI,Hm1E ed현<(#N1D];Ӳ)_Ҳ*s(}_ Es bF3悡U^*-фE0Zέp :Q;TmLI1 Qi\XP HsT"C4X a YX3 J괱Qc9g(b:E&u`8,i@8C aqu)1zɺdzYɭң18cl8.%yѴ27>Ы-jl EjT OfVbFgV[0kJZg<8CI`s:KIV e+%|2(r7_'}>U0yw;,{j%&e4SN[! ~;4B{ɁsgeW+i:Pj%M_)8Cj\}fn1nWrҋ[\wXra&Yj5HwfXޢjnWa!%0 9w1̚e xe[ dǖ= ƾ|*B!d(xWMH wH﬐S;V  R{5wu!"JrƋ5TwXq%6;mE LhK L3[1}ܺB8~Uo~<=hLfJA (҉s1&bKw3y4}sT?Zg i27L$¼2,* 9%;mٌ/|ùnʭkx U"([s>!pUQaLo>ᴸf{{_ ans-z(Br) *9|2H-mv|S> NNhNiDъo!H#@H8Πu$m [RD*pDIUyCaBQhZhDUmyxdPH",F*ZY5KF~^FeՋEk#ROdX |w :Ѱg}3:/vq*eGCzn,U/ )DǕE muyj ̈BBfn0'2s7,t]ikY qc,d:scpJae1e?'EINS|gpcKQ|.kڊK-Jxqg|Z[OY5X3oHw \|*/耧n_spK66ܹaZ_p. 2:R/0/(聅IJњ&5== 8^u 7xXm=7n{V)$oz<1d:WFN9 N#nV%>n-cnמnQ9u^SѲ ,9-T1wO抐zhP>5׹1YF:w6҂O หUrf@:K3אּPҺ׏X޻^)N4ЊU3֦bqd3oXs`Нm]XTN~Tc/9"Peԝ ?/lEmûlK}!pzm%=?{n^8>T/C- s;Zy耭9ݭV*iyIn GGq4ZGPQD0jH7/nϿO 48} )EPVFyu;6̵l܆:chpz^ W& ̇Tsו5 ms 1GV8~c7r>2匭ZJL L8iX|-CVPU.`uΘ#9RvTIb:S>UAu;ÌwUbHJ$JO^,cӖ ]6!m.k[H@],nhq<WR7䖙#Uܘy8&x/>j{U]i ɣ"{ c4 ii0PimpH@ R&$ĥ`a A`+-Tc0.Nlռx`nAfo4]"L2κe%b(^~od扂#Jm|Oשjco4~}Nx7Q-=5%p1<90 Wʊ _/ M+b%3a30§ 5 &=T -eDmN(%!=$ 76؈؅h@Y}+_4ܥ<*.jSjc?wcM?xtY\ 6Z 7!^C[?kq#ir8j/o$1C0 &ZDCY ꢡ$QE#j+䌲:!V>RPlo)k` I1Y4'#Y_ n".y?4~NNN7>x4Jr};g2WDptbnƇ$ =`+ W0]hX+q"Y_ռK2c̎VjeܶvYWwUnLSjrVUfDd/"3#0Bғ{61dR.v M,nnœ9R:+q@ , 6O =~'"(-a@pnH8U1E[xVuΎ]/lb?f@ o!_LJ%&{ڄ+ohiC (6kt" DZst/;#cH(]Sd*"\uIz))gרּu͍* ѥå{7sw?/mh<_ZPuLySuӁߔJNAZ ;V{K|%ԍKwȼLry.MDͧR 7`Jge,6Ye? >iy%QU $QgND o,>c6"gQ#Bp%@S޲1v 8e<3 E(並Z[% +j$cfN4G:|,LP} gZ|L\2mm,S%oNSxi/lL{S|,"y?l[6.-?n=esr~d>9^>~N ѸݚY{ʻ5`m*9~)f^+ E+oIn"ЧXB@[[y=kt3ML[ 9ԟ+JYz4IL*ֹz־bXUῤz@L—S`"c{ .Y^gin;hF94v2"k'O<3Yߝ<2 + sQkq}|6wgv=׷ ;?"Kg>y?4|=yֻ8j>_v>x}y޹=7֮?E|zq~z2L J_?|S$f{~jM/^pe^\:'{g7+pΕ?76YܤS%g{?xq.WRV4]﨣_OŘ=/X\'\\z>P|{^f05'wl-4-fn /ݎ~ҽ6?qFZ|*(߾%?.2 W]rc??:z*œM^9?%,1#K|}vȦz Z^Vy,k|s@{fя]c M,y̳<8 *s/wȀ" |T0?^X+(UHI8 pp)~AJNԞF;|B,Q!sq|p>Zp3"#o&Iyh[.Mͳ}L BG*9/2]9/2/u=Ry4Lˤ2vp#>leT i2fP0`dƗ.|*W4̏}ˤL_+B0/eG}!nXΓ?]2C֢gs#rQ#aSm6WP5jќOgȊ^ĩBNC 6k8UM#9 Tpc WZvkSRx$CVJx)qmܢ=+z.Wp92;YS')`m$mYހ;2Xy lk$3VkΚ* S;Ƭ0eo2O0ʐw8{'ЗsɆ4[.Be;Yd/4aR&x Pű%j"r c}EݝbukZ:[}C?=TgRnټqKU'fʹރ-xMgvKQBB@T_c,Jųe^h596o/B%~r !}C c#AiI;|P[ l5# l使\j[M`kk`k]%mn<6d`m`èUCdk#(ml"[ (UxX2E^Vy],vx%s[I1?'g@(0}jg(Rݽ7 $XHQ.N]ǣV(',ڂYRA&F0cy^HT-DzB]0g=?YvÎ퉏/WM. 4C!d\9 p&)DM9 q ` .9tsD2`N`k#Ar [RHvb2]Q™SS3⨑JPk"06fXP8ޖz Jx:5*9*1ztV(GHk0Xy NZ. 9P{pz"pC5*>ʆH"UY]T0BSޘmpY5,f1:%8& fnQ|ѺoBnjVt\&,&i+Q,FQX5.TE%JIpЊ4 d1,Wrx%hFNJX-@ep IX"08P<7y,(/F)T53~W TXSHgK>"*PT[ZMLϻf{y\U8 q̋01Xnah[*/fT0y/ ^?sF T+D f(e К:#r,׌ ֡`7LIB< -:>V *IL[N]XA|ynݖ0gwmxqJ)}HHބFEG6zKmRicqA1 czaN T`9B vH1ZG0Ӝ-"UX+pwH)B*V :?|GJrF*V5yk"Ty6+H1z"G+j8SLʱ촮,Սml\ָJ[ǐrڈezۤJX}o?>ƅjZhJ\8X~ "e d"$g=UT`Xkk|.S˽$L ٸ줧K1"6؅XK뤽3['މBi9u +R>hgC45Ӡ=`l [nC1 ,!&%aXuvADMʐJ!$DeF$QPn"XQq2Py28O[1:^U*myk#QX\_`O鹼3Sq~=wOtx׭~?{?@>IzP4H\4!Tzw<CiZ:VU|hj˔Hf8g:8T [J E毪yb2*b?؋.W~l&@03]3`2Y1,t[nB4w {=zWPd|z/q.",&OE:QiKl=z_XP Su11QM׶FnW+A<#κU7*@ozFYY~fe+`7Y `oˍhI7+kjL_-_.#)C<۫L.<xe)VͶ[SJ^h-}|q <O?95waa?kP wu4 ?2qӸngh"A)ב iuZ՟~2m9 BWG'0*yŦY-^L34߆e9#$@(UӄC+.$mZ}p;$悢Npz7j ^0*ݹϿ~ n]^R.Ϙ)bhb &gϋ;|xp\]\}vDŽt}C$ncC|95ߏ9h(=\ >C,j1띁 |K1)y{u}juku˧:B9PCv$d*:U)nDBubyxF*Ќ#.IwA ǫ A YxxM%ST.UDAJci J4hu!d*f#9uG{ĮG{Įb{%1!j(MhlƩ2 P&k)H /z_ZEv~4 J$G59keKq lQc̷2/=:\ _`4Ξ]! !YXꊝCiЭY',h1R]5õ̐z~ ~u@f[xd ~3-3͗ /-l=/Z6+6,'ʮ j%vͼffFNX5{M ۿSOhN4䖡(~] l-VY+"y.O_B'RmT\&DJ3f 6ibS"Mep,10cdMcbL3{+ uҟw:UD9_ˡ yfG[2[܏`RS8lvQ IV)Cފ rɘr bFj0(n Qd<}pw]CէjDp*nN & 7?Ӱf(j)&Fn}R] y}^_.n~]Q {{Zw_z\gΗ~:.{ e/2ho|y_?Ae|6Mv`sn-ZO퇹JZ̖;򡴰Vgv&pS&Sx蕖jjtS&u$77~f4[r[1V&tp!. R@J>?qb}~ha1: ӧ`f _;x̋cIY~֔w1>LfgX8V~#\HjX(x4J@mtz8+.=5ُ&S3dAp˿jlefV$瑙~o[M<{QlܿT vT-S`GitČe3GHN7x(wv4,lw|#RȎ 0m"!HJ}>NY9I˳]to\ȏ8i H}TކI-}#_CQ| kZ ri+| X/Oŕ=O_ '|0@ b)C)cAüՄjxՓ0?*N;qgUOQKb7jUz!zRH,o%^ 7PsTGib*+C tG T GW/ex(ri&,T8yΙ0sURB6))pwΓ K'\=p-5T}Vc'IDWGi!` vS!!8}DNuzn*8ZPVAhVWgmI ;N-,ceC5+;;$ \#pwN-`E޺s* #,:rV6w*G#)ݣB)gZ'q+EG(!mѝ]u2홺_VՂP/ u<2d>^DNA { lKTHv,&f%8OFpK]\kHqKUie皑qETU_}&B*` WKBW&#'C @ȡUW@-[R W0 vrjZN< GݴnxP}&h.;~dD[5v-^#/pgA76oݮzL<bhIk~4 i^O P=^])BlkHwX)ܷrm[Rvĵ*,U@umK>L5?h{>}T'!?Cg=C0oH:o?0Fkw5h oas JBw|;at%up5|}X·GqDD2 z$?4)MA?8>#80G\=U{};ҜsJ{^oE鷐/X uMlL_ nҏŜsrס]e,G*Y>)Go{@ug@]EN*Z~N [AY_m1j<ĤнV!΃ZŨt% k-ᨓ]mt|ݶU8$oyr1Ul^]8U`^]6m?LZTOOq-UomW<RDsW,0wa:1,qQGL$4Jl"吲(f1zޯEmm='HT%Ҳ%\1PbAb`Y!cENqc *U[W[5 ^fqGn "'-`0($' C0ĂXYb䬀P9 K|KfWw10$T^Rhu8UF('r8:FdM ^*6BVE,^}fmt%G͉dT Iaa d#iSGV֤Rjdp%1Cr$8U8w $@ʔsXM*V@.pSZAYH.5ָ@QAQ3JZ"a\jÐFSMM "唧(lKkxTXs2IM˺ OF GT2ʂ#k%2M\*N:#EQPY0r`*R"XA4qZFItji0K0j4`r)ffz%+ ~Li[ĒXb֫9i_*vdtƮ\VSd4Ѕм 8O73InQ :0u4Vkh\%ehtWѻTaTهr=80tٯtFsH{o;4hF ^jtE _l{,'Lp)KI5$\H !rJR;/f_i@wS*A ͻҿo;V-b0d Teܧ٫( NTD 3" Sl?їUv :彩2>=NeڵaZm;Z^ԥUN}CkG$l:AXKiVxzC9ݭV$s25j,Li"S&X敷*:TX=pxAMĪʩ"uub"X}c$bF7v8'OA@oUR"sE|&.&-NFyэG~6;hcr8?:tC(9?_tnx҅D&_ZP'T|#˕ϋo'(6&i+ָ|\٥ jy>R?YX>b.}7Z%.K"v6IZ/󷔞(/~Qrv$.2ٗ+#ԲRJ 7ɗpTW"?6?$Z FLhC-S!b/~-TC3a6$ʶQplP2`;]⟢g;(EgU'|*HʉW~h_z֔x77E]+]ٶoQ.p|;\}iʝ"t ^iN\1g$<ſ3Ѡ 6($-/ؗZԵa}NQ>u|:cJBoBg0ep cs tJ ϨwCճު18VxM\aǨ2Q8HAj||P6J>wA]k/xGۅyQ.+'MwFw8}z撽, U3?߻_\-wS3u 4ʧ׋2Mލ6(@?0 nidS<'/{~,|(/> |I:ؽ}=km=W׋e9B6dhDZ61Q1 $6:! -]؋Z3(zq;b.ۜ.FwnCC1ݎ63y]'2!FN|ujwuC"[(Ny΀Le:tWre' \\a$޹Xx` jM\߉у1[ىMэA譎mCz̻h<=Uh ܥy?K!]8jXf Jc0EqQRp.C [r1KCex6,yW~zw]EV>{ΡX;C/<~~>~`{⣟&;/қ;;=ף?ލQ\ﯯ1?4)9+:/~X)3]q-|X0G7O@T0,R?TPN#.xWJ={Ljc>|CZ59jQqZj Ki0q\1Ve _3 ,u[>Nƞ0E2úJX|mP%h1Sĸ)Uqw`UHWj{Do+(`RȺz"m %*d^[kZIh*ƪHp1ZM#N4U I'cLFFI۱kh6%֔>N'էNW :=G|'A<I|l_ -/ / @a`GнVZm^@jhIio3ILi%Jڤ# wq:1Xq +NO(rXVdF`vk[qI ޶F@WDӨdܖΥ$Lal({zB㜽^ KoH>[gu*SƘ?f IRxp#5!JxQvjf!OR&1FKAÙ^QpQl[:ގ8hXN, MK憤qZA@%BR'S&S[lfLy-t3 RezvbE}\jduΜҊ|R6rN4W~Vƛ}[008 noqWxӒ2C7<IO0֒A8(z)(>DhmCjObt"9:[8aAQ#{z&҆FH4c{FcU0#Qخ3>Qn O6>'RazLjDSD*\9H)pqDЧ RY!'\dj({xPL,b*,tzMnoyѲi5%gI TRqO_yBVW{doCCơeug9Zə#1A`*QI&($NSVQ=OQBC%PE6UB >·m!+ N YφLpw$'{,{oEӰYηvP+_u2B~Lyvw+Ĕ'u'l5>:@3EyIXr0ooЬ{NJ.FOֻ>3M^\MG&qgܣ pհ!:!L dizC9ĶO􆘰 v ÕSTwڽA<q麅Snl0e]B {m,59T}yCo1fzGoWIzu(m = FCn<4Mgq%AH1x '~w ԴP]6Ǎ5TP1.j K-:4lt XR5HlHRrg;lGxOCKm^ 醚P =Fm8i$F, yMka3ܘ64]r[gXS Ӕ卥Y,S$YN$sRhfɶ'4``D ɻ 6:[K[/L#XϴCT:-5TA{?̀sBhjI{? 'l>N0N癦y[L)5VS14gˆ4e @<Wp_jb ʍ!:)g0vsR9ԑf 4tT0KX+GbVOA-! T LQ u&+♨$ސ'QSD*v{f8O8pBJmukr;h/^*%㻉s?f!9 }d@=l )ʏ5 2sTktZ!??qxM)F46 ]1(} %w؂qB+%L 9'^ NEڡj,XwNE)X+o[w7.;Ң U;c3IɅcy] 娫,Fiꌒ'}zDܟνd>1a\FJgFRaxA:ʩCeܷ} #&sL;0yt|%MuJd %(gf Hrg~MF?4u*aj=LNXz\q+%F>@U "wv5tt.VJjnX-o_=ᯅ?w֤"HV卿YYH}@wK'_no07%1Ch Ez ~wc}4]t!8-‐,F"&(e(- (7GVjGg xI<5}+@ ĂogڟƑr]Ԛ2f\vgRS)c]dI+QN2נ~EHUĖitFP9FPF50ƥEA +ZE>u+RA;8YF3[,xN/)VRbPB]/|dXU -K#-x?gJDrT䈲یbkқW>Ȯ]FEUFw^hrXvl'P_"F!ZK vm Z1dͦDn:~H_E_$\1E_"1k }姏,<?/R.0\jtE{ r8 FZ5~S{ouc񀺛1_?&RVG)i2@^lGA!,pB}[N\ _wh{ƃm^"=ab*[#Ɓ:5[Sb9w6 ih>%qS RKD(6ϵ16/xʢl1ʘ\hfS.8.LO'ɏ5Kd:hVh׆ɏu&!vZ2I6 ?TuIua^E΍<ҵ&=k?o''54JLQ=)F`!}糞]lZ]|1XנLc2/ܻhF1+RϝH ) ޺Ȁ ? ;go$C3g@ V^ؒc!HI z` :"WrUy8[ܠWC.ONO.jǓY5YGTƂ1~qvf@aW xc}qb&4EB#2Rt%[;z7Eu o@ `}/}:_@i⡐0 0'F TE85Y*Q.˒hz9 Z_b9,y03 Hc_ "/JQSFiQ 8cP.ǝ5uC""qqhq踸 d\Z%(.A/Cf̼y.vA!&%4YN50~F  \)2(R2SRj&฀mAE~0H؃!" VH7q Wtl$lrc6WwV]<99ʖgFu@e>R .Bܠ }΃r>>8)ClԥPv#IVjD=޽@}7΍}:8sL;WI_)!D(F@Oڀ!'RFz2=i#:n jg<zDz `@²2Wl%?Wކ+@4x~ߝV0-D\CZu3Htw)BE@Bjc);)R\G#W8̋bQPf!EB*cJ@TH!#U W#UB~0F+ T%$3G8UQ5FUQ\ĨJHTEI"U< p܋hr pŏ᯳S5 vύCwtr|?t`\WU4Ue1AD8lnOk"?F+)ϼ<Ƨu lzɩ vh9k=ǠƩWv1_vS&Z!׿{1`ms1#hЌzܭtTRN%Oa4˅H23r&&_&}VpoM[I6ojߡ6ph ZmC)힣 NJ\]K9AjXש)t: u u8uT*l K4Ek+KU"nGnr08Dѓ Ɠ傐#/pO@@$7rd[11_ߞ=BH/u#Eت%/ I e1/3rDBplNY!Pĺuu % b>UJ|ܰG_\`i#„B fBц+)ά0dOS,c"IR+@V&I:S@zdA~õ1>{| [?GDp q**{ PE)b9GGwt*{.RpJh\Y>^qe\[@o^#VW[/ۇo0_% i 5*ɽ%_x/hc7/7HYZxy,_4Յ>6c }k٥+a|6h xs~k[߸-g:.*u}szoj/|/]wųޏtR,kޫ:)@G?llvEaR?zE|r [l`k=ݗV{Q-l#➜&|;1d5< *fRfbi㝟fiݏva}%M3..sW3ۚfȽl`8%30G[ܚ>fVnĮpzSҲ= 嶨g}!ohoJc*A*!D7ߣVCG3ԷZVFV0z?V󙣓f`& hz}2EuѦuC+:]1b|^\&yf&} ;@ƁCUл٫HwHW5 7idsIzZ<>4 6ܒvKZ62t1gG a65RR#"fs8~8pδ8v8_xPC9QĒ# a@xBx$%v,.^0:֜x,xvDxno|QBf{ xD#(QZ1\5޲4 J7(S2o5j50Noy~3N앝Uhe`_?px?vzK&ɳcXMrT) F VY]U\Js-8h!-%Vk.DaWבW)iEPuټfV7՝Ygd1fJ̆SUلTyێ1p0u&Q@1&w (fÙ$1fP̆3MX Sc6!l84&98fWm =)4δe)–^Kp K"T[6ʙkI8ײ䥀 B-3rDaaNCKes !"{VR2% A%J+]TF F7)"DE@-m (^w;+ kAȱ b[sDtqUtq#+=ZF7!7`<& ̸QFS787Ä&Hm!ZL8F  WRrYyɐUH SFt5UYM2dF8/%t!N1HeB`Ÿ`K)۽B5 ׫y=u55 U6leCzߡL|i&:荪sjtҺ_e_VG4QSh`~,4E UQS!ŏj ;Mj*YogN#N2Nxjݾ'YUE.F=;rۈ7wU:MNSg|ߚG2l t:\J3DV61%v<u7zG;Z@I;Z0?T ,&eVIYaR,v{G❸8Q`P0OPx=l=A#mfN6^+@qmNFYwT7qwK6OkyCA-dkRK {j9RK ,TO-gjƴSԒe(FQ^k^؞[֤a'W8+FJ+zogc;0ХcA=/P/&)jL> !}5S' 4Ki|BYw}}ye,PL|XU3yB4>ळBƝT ҈)W@]oA [Oo$ ny~ɶYf^oiQGS:w*@1b|;u6?8x>$fp3 ˸x̟14DzV']X~Q>Yoa)[Msd^vmVOno/YFGRFX0.id`p VKL49 l~3j>i,Z?o}n!~0i La?'t6Oo)x[Q4>%vWEb6oLӷݸ\d;"䷽'~ JЛDK~n>{SqE>u@[±3@UUhҭAsshpiQ!5k؈Z7!/f~IɹSArn\h{l"/ {3o @jk{HkA1?8䗢g:~4Ƶ;x_ܸauPZCӾ>9_LU O bl_]0Տ>IcCO~K~yO~K~yk-AFs`  tc-;8XF+KagOx<4 c4-!+~lb3IuW-&h} C}0n$` v%=[%_|CZ|$~i'ؿJc xDE,:[AR4pr:y9WO Cx)qF&IX-)VR`pRL򡎹N1ΨJu™Fi-d iO@šaFXRD 'J0=dFaTM]v?՛CfeVm+A uY=s|g'i_.Z5CK]~tJspy{[N[ oY e[ ܪVX5Ivb Ʃ[ek:9kHQ D0J;M~}xFcW*pg8O^:Lu>A§9!${' j'Lz|rg7C%7=rÂ0GgH%eC_vgag> >u@ً /C/X㗡^WE!ءy1RO Z_U__t"G6(8YGЌ4K HJ ,XmLf _TkKb$\'F)s8u'"D5^X 7!:nчM}tcsY|0 b%-]Yn nb찺>גX_KVHX_X;K>4d{X8:tqBʸ'R )J)8 NHS2T|gOŊ@2w MVg߿<\Tmמּ;:I*I/\N5ym3loocJ(KW8\GZʈ8J(Sm8וĐ [|5n>doQ,w_Ot܂Fgf{lG3?-]C)"Db4<|gKCϋQ>8ÐٰN"E?%l Bٰ0/Άrٰva^|ߺpV/HyQ}"5h"Շsgӌ!TЎPA;VOIz{Lg0z8wzs7;_7w/e //E T_r _oA/.^PՅi+-N4D|j2nLkNK^&0h)$HA֌pIp"b0S)CiX$a"*Q`S)y,Fhhh*WfB HH2!3 דPe;Me314(VҤR,c-bmDMkBsdB47Y-ȧbxg ZXOz} l&ݬN1rͨG|Ȉ Abw8A`y{SXq Vd&\a5ŊqiD 6؈HD1f4 %d`zАf"&)1V"F KicfcZkb%8aP97|_"3%eu~$PyqkKЃ_d(BpϤL1g%@4N&(bHiXi)Jvh G |ljI"Y b)Gq ##Trc% jE>.D~E~Hal> wFl60MJ3,NMeePKNxK[cTE:m(Y @K$-o}՝φ%x5tL̆,28JKŴx h{ɪ)HeY9IwK2Td8Nhd0ͣ$c41:Mu$JHe-, e, ܎RL #V ')\6FF [y# Bws ׯyo~a&xz@ Yyk& #'܌9:y21ڻ1QLЮ"pRx,Me8GH;"Yl BAo)Ū%VGblWSٰ͹[5xH8ys;\Qck~)e$4X/3\Kf ܨweeҪJ*:7*<IBX'U%YYE/OHT!#l*ˆгk%,AʺK\;BGihYTRk;fxG \^)*2rݮx@n Glc`_^ȹ^+Zŧq!byq|5t!uXtΔze6q)cPJW  *3[M_?idfX>-.jz>3 {|uiEZ=9Z魸gD+4fYLԔէ.Vj.󩩉jj-湐:FO;xAc|`CQгᲰtj7x5OrL`J v\NM7mR3F.}Z&lt:g k[g|\)ƌ+(B>=g +Ҟ(iFCifƻ uw8DHVo62?+SXl"p9ZLѕ޽1|7TN)tnj#*usTsq` Q򼡵7/wT8fBY+Hd@PF@C+u Co+@fy$Ez8bp\h3?v[>LP)إatߧ_D\J^Ǻٞce]u"m.gZ'j'>֋zbǂGS(#ǼIUյ|p@O2}jۻ֐H##R!eQUli.xDteZ'R#\9yT;1OJ%\!X5O&ө{Ӳ!泲|堙rUf>n=9i#w|R*>gJPG e1|*P:nugxJAfҞD`vzrx|^`OϮc%=M)WM;\X:`ήQ8S#JC8ya(3'!dL_BHtMN EkrWR3GwRpq@ذH3{dP SM_/8djixOi dx RU$rEɈ Իz4ʣY 6Μ9e48Kr5a*kt:ıbˀ.CB%u{"~%SuQTG)οi[$9fkIӵq`W~(CY_ĕId7vd/OV7$ !JZT[y0l"S b;.?o~y͛.lioXMhV돷g(w+& Xnzmw>RMׇSb%uf 36MZ1W* U*9?( T?ÌRбxOL1fq?`YINR5~zJ8J! B"ɱMZ?}JGZNNBBtoT uۃH)w񝷗~N&*[w3٪E(U-`< 8Í6޺A5Uc!E)I>:jT-p)+`)O(=)k t{ڡ2&Kh#C}OQ!e9^ɒ%G-y͝j3R׻{9+Rm(mNz/\ :7}0fy]r%66~ebY`轚G3~PןӕGmwSZ(+Y` bp fx.lg`ΛnQG-l9:arUr`7FN0q2Ŭ,m|b14E߂Y{9.˛ pnޚ rDݨ[L36Nb^[NO@pi饛@eE>%|hGh``4J7J1qO'P *Ai 0o1S: u1ʶ"*ѿ_IYhzЪN?m$cc$\oO+>4d!D ԯX<r֋`e{PŭTO[OY{_ꃲJX3sg 5DxY7q8"0N`W/I3ȼ'<Ҵ|3ƖxRLEe~w⩷V,+0˛&l+1oDQjB)0##Z!x^C-k'o̕˵ު[1;F?~Ļ~,92onӃ:c# -Iw.eJR>3G5S5R3tT*ʍd|9Z>G orx6b?S2+6T0Lt}D 6_nxف䙟^3lʾl[^"1Nͣ b>Nˇ %pGJG/cgDA)j"i64ϼ`el]ɧ7_Fr'v \5*ǽ&Q·:^)yWx`|K/O`TW6b5J.='5scNycj ڹ3.<,bF=Y9i{EgL>,;e%:^qanm_u;zm8o}GOBߑuy#ߑ= 4&s hZz?qPY.Lgܽ \wuê(Ŗ3n>)7۟*>WGyS24 y}0jUfw\ WGD0Υ`b1Bz%nz8m IPjDOB@%HVZ/P]޸,0JyEB48 LjH/K8C2MSD#hqڜ @آ$? 2 Ey9j V ̰[jqIGJƠ@>QɗJ}v$L!PP%-L%b\$ qadD1"@J xTcPBQJHD aښ1e׸)&"?J ǃT@@2 ^Q46W#25]SG춢婙ݍgmra]6DzhA4 &bՆޑ1+žƤ/c hv;tIjf۾Haq}ZeLɫaڔԛ̦\\Df O2i1`~4e)lU>G17 (n϶l>O]̈N~D)+Rߙ-̆ӷyO"!'žE)(bހ{0DZ>2쬼Aq]pP̯g AuV'u*<ǨFoc}563$./JT"-iͩ&#}/]T>0m3p޲oɄ;KF[V^4QQ/?OXɤvG؎q$'Rº\U.ڵJZ]|(Y},c,a%nq$/[ZpwyW,,Xum? LBCUm-yV~u,,G7Ti~GQE3ox*01 0Ч-uJԼȯ0S"֘]Lҁ536ܑzd]  LRSu~`S,EK'[i?fR"IIUoǀrP azYES!(bD AF$ J4ʨJE4QqB4R[~Al/52lߥdPk&\Krq*BKFBԽQ1%:>?ϿR{ϊq$Vd.;_"uww`NQ &. ] 3tvëŝeqVN|j5SIXyd@e5_)ZOTbLvoY;b9-pTO ~mc1>b8Y4]ޯl""PP20ӟtn+r׸R |R|"dB`%c#>H#\&dlosޫ'Ku#{KZeCݒbʹp)adm2|E_/Wʊ:q,T̾f˚ TcRYu |UƘ~b~'k[eif`_6@ ]͌&N.̗Ob'F]l -]/;x-B*a apΨ |@.&(x&Hݕz$ ^ZI¢:pC"F!뼬bھZedE9jS∳TJՓAirFYǬomDH~ -D- /6 Eז ?b? q_$!rQ`B<DE3QP K/h$LH}!JdDhqyP)Z(qLcZUĊc.qڱX suvbQ/,$rQsD" ]MK+VxwYg1+Tg|\s^N-]I^6&fJ}(BO1u3~XO0JEjx۞D%/ ufwKd~Df"I)轵At{W>Xt1-Tz򎀺X/=>FpL\u,/9!pBHX#20DiޝGz)qѯt n߿=,SVe.QH.щW/gQPD#E[bG A\|w 2R]LT]"E%5gvAj^tv7@9"RNhٲ} ,gaWiڌf<K { "BJkedR΄k(͸l" z)ګxrk"5E'%F1i7ry 9'bn >N]=*RJRwjܧr71X&(ñ>(3zGrjvN`e.wWO?Wyr f˗]/`|;G?Hg? 8(qY/vOAC>0z]>6u7b2On?5HWR\曗{SofO!w6ܤr'zd7Ӊ } jRt Z2q"> ͉'BhQAih=>MpW=x:Do~??ijR^x`V/,KҕZ<ԟ?GxATK-9Wn[1AKRBڂ=X9(`Sz+ZuD"=K bd D0J-2L6HXRn#wђXK 4XX̏԰l֐BtY."U*mewL, C-QdywZI4R,iIT2 m$0IWN 5J`5ylXk@KL:wVf):F,uZ`Cumކ]7Oܺ8b7Kw?sG0k"/}EsM_G?N]eywE k35sxo?~qqa={w$3B6SY&ˇoz6- z"~ݟ)טV/Ŭ҇߹ҡwr 4m fRĬ~rD`2|DJ1…".bBUg&s)FrWTDka^\y)b.Sr)5( r ~1u"1Ddu-=v*t 5dgX2Ix>n@W9)T+ZʑE,5ާHz ~wJuf=Vo:"S˪)14z%#)(q>|w ړ/`yYtv,OϯS])U&vgoGM` =9zuw_~zXnon0~[%yfzeim̧Q%b'OYb׻LV/>˜a]=3|d[Д#h9qXr 뜙 gt7o&_>"r/2'{I(Um , fEIn|9ez=~LZ x2Fq7l]MXg*ꎩ;;H ՝Wpf > Y|9evqKN|O `J-' FV_lwףdV|WofnkxZL Qǩ$ )ݯ-YD 3)bh̑a&su1v ޒ f'0c>~㏂Hln(0pza6qV(Hx*vMjmE1#{~xwc\nܑ5[7sԠzwvkۂv-[L$]F0(xx DRY'WzV!T"wC ->Dj 9ZGރMQ{3? %.KÀF?NΡ]bZ(ۡ *嚐i*g=SUH0ixszad8.!c/ 3pO,/'|)=^Z,"4K(3.Yݫ:9( 5fB_I z:t.ן ws9La~`)xvx'8@i}⦀1޺v`Ed~f;&^T:39S8=Zs,1vCJLW과% RJU=}i0k)?JAμKۭW@stȥ*Ђi6 H  6-TaWt RG @C[zmb~! VR뼔lP|/(U+l-h =K8.9,:$,R g"u6XXH#Cl\ܢoG_-&ۮItǼ!j]|>li*ӭn!̉kn@|>~#BIDk"MW F^D`wr T]Ñ| ߨaϸc|Ƽ`ϊo]oKo49I$y:xTh5 ).a/o|gϳa U*Q +:*T)qAD=̺\?\l] olgN5 uЬ7WvWj|l\qJkޓpq4Fُ>UZN#Uy5 {\/dO> sk&OĢ@gF.HBNW">S ęOBN'yͼW9' f_$հ+Ew 7|SG_Vx%OZbJprI >ctA()h1HO7q/H)Sbvm^Q:XX 8Š9a!4; m_Żl;8vSo嘜eCF`RuՋs\6gvVdy 3|Oc, 4s)`S+iJm*he#Ld 3ε^Z.VL2;w5`VWrZܨQ]=3`oa[мfcfJ*a8M,laH2a؟w3y3r=$6\6, Xm-޳~44tgկ/,dT/)'3./Iv`3`(ҩM5xy@?%̻$M3Whh'<βwdB4SJሷ>#i%5jJǂ"EL(Et2 CNY"ήQI;XMw%ϘezHS%z4KV&lӄ1%`)L WԄ \ 5rae9lYʙ(qx^4?iKJB5a. 16%@J*p9~Yɑ1c1(9iSw.d5y|?>$"ͥfeU {٬%px]2ƐYNmY+%jvYJӕTc;uKH^N j&O ol(I1G)7h?xwO?E,k-tJ] %0Dm^=N)l6~E&H&k,1 „\M< oo$۽)f,ƹqdp5a-V7%Ep//I\m˾$j7y櫯î2z |5~ñ3&io4y 62o`z†H/3Phn_޹¹Mӿw<v|oғȽՒ$.d[BڍʡP5A4v;u+(Z5&4V!!pm$S% yv A5A4v;AiƄj6$.dnCg-%f-O5 k&&S.6@C9- Q8{M֯+͚8y<%.c;|4ˀ!Iô6PZ*wmmJ{v$/CٗJdY0,n(!)V )ix03eڮt !i泦JJ޹}@ז1ʥ;ؿV}Pٿ'M/|"Fң$ڍڭ-bD;hἃS{n n)$62Ey$֖1uېQwkMhvK!!߸ɔ)qI@r-ݘeJKCB x a|`YqF^pi(>rᱧHk™2PcR{y@밗CF\^":hJqS*}%Q3o t̗Ezv(xAma7Wfy5ګ *`ھ\SbyB{⋔dmkL ǂҀO閝(ƽܚ` :MUM[\66 q2tăّ(ֽ&E\DzRW]I6&cXx(BUMc2K)R<)ە62qb3C5EZKcJv,yFq۲+eOJ(٬po~_/Qj;ƬL8acj%G}Q9GcbmaVJzmi1.-T\mE5PuҘЪAъ 5hE %ZVk;oܬF7ڌ}F&^ueؖyft4@bg[&qFݶlU[x$XJT]z k6Bp=՝2k)s =2A!P? Tb-F49Ex8@HSa_13+z,8sv&g1ɍ$ m B.pxp2Jp!\^a8)b#=WIFJ궇-f-4R_w N㠱jSn-.z*±roxrC|\iCj%a4P*u7$ n $<h n Zp)s)b90b JW!r L$iH~I@3]eM:]I(naq' Q̄`,(+kD启֬Mn*+ 4fZsS(r*DN9v9oq ^S ᬃlq,U8iP %4a &g#g˜cm;vSiuXDž/;9TZrLӘ#YO甤laC3/D0E+ ,xq 0Ӕ"i"\_pQ+.A,[/^Q=h6q;1֖e͙ rMKjRLP%PNvd,X[V71@9Md2艫wZK8:ZI쉐|q_ݹb=Lϊ;~g5SV `bo說k~-"$$R9oI$BBH IͻCwE칔(fNX*MXtAk mz&$~CAxp9(W۔f=MIiISz$D!IH4̧'аG#ҞTćOBIRH-Ic)Ԛ1&QnB h9vKŐ.֎ࣉSbT!$ʥ0KTA)2c#k 4otBjve~b<ѕs%6JC(478PR_~.KZ)Тp妡.7&MIw}j7IyoFݚUQ0K·q:dj-R25&〩1_05nxK ]|6.Mj>?wԬlfb98YPhIho]|o|Oc̺Xŭk>e_f_DZyc殢xL܍Rǹ;FuɋMWy㦌;݁iĶX.жXǞ^Rױ'_:V-I*jq8m; HGK*` U.)a8&>~YƻQ؞xLj͹Pq,:9r9 FQ66W4D;]GU4|yzӾ44JYI)6f1?:vb^U!oIw|{Ȓ ~m+[{{vB]ݟ(|; RMZ7`O^ z,a)c;S&t|)N~ͣNW|%P;&H+orD_{w|અ(F+.y%cОo 03! R(09x#824 O溜+pEwS0(IDX:R*abKG~qf.hEŸlp'~qZ'o?=暖S"zanA?Է `Ue __bz~//V B?jlߙo;_L`n(ӗ'a] i,/6g?E?I,* l_HKL"[q~Pj.o 1gX&d -aY1PA+/ Y O &ϷkTM+⾰9ôxb[ЏY")WeKna}<&|nET sZIo3e=Mܼpx)]6T~;-Qs|N Unal}VڽOp~'߽7i6fNn&5JvCeufbPf0Q'a"dRļh&EYkzlxм> Ӎr0~Hjo+Fݴ)ìηCd&MiY xHR$<{"GүM:l9}v(6۳CNNg!lb<|y5*wͤ*$z5LEE zst !юK^A*XY}t2 3XtٱQH=GQbcj95끶]u ^<xlH3g3ACQ^iT awA^!+zմ\}.˛j][/'eoFzTZʫ>znz?e γ* ^6!=|fxe| >,Jk Ugyc58vxIthd]sH]YsG琷c(u"KH"18g#"+@t6> HJz9{ Ί?{q"XKqyϭDe 罶v"yܩ{:} J?΁Spdu.)jN|7y>FfcSiUx q5U9a!3k,;M7Rz1~*m`#)o%C(8 *#9Ⴍ?eZUuЌNK*ɱR R̓*=^JQ[JQіRbK kĚZl</L1%F,m 8I*%{)Fw~{{7]W|Bbl >!]YoG+ fE}Ѓai`km $fӒ5j]$ȬΪ"E=Hb2#ȸ22rBɰȇ֙1,0az w]Z0\Q\ W޳@%2F ꯥ}@n1.{l%aE88jdH IQ )<&XloEǚw(6"[E^@DIA;0U* a$md`"DAD#wDZkyolgHì\Im\ $})X )DxEΝ TZ ]Cx#4D$n)NޘKS<T+ZʑE̙FLO0L #AO̽K\`OjZISʰS$%X^4i1-Ռp"%ԥUX \t8r%Dշ~Z$B|rbB#AdqR$"i5<$N@X[b*g58\Wp[ Szu`Aup`/@3jcHdav&`[Gf?'j>hg 8?3>5&3kP~@ǜ-O4`S%_GzTwJǛrZH_)ߛ_/|xJX)^+o^VWi7K,H6NHQ:>~0RDy=1sbT1"#F6jU\ы~Ũe91*fpشۮWS{#8Ԧ=ڪ'CY{NYY"C7,Z rh/["|H$!o$>#b $q #4:'xc.am2 ݙW]pu} X`78[>ktTV<:뙊|<\0'|<,aD4ã\|_ TEadͥM滑Jd{5oڹ|ajR63nMQn;zng# [9jl=Y5҃qK1&r0E),ʸtN큤D%x,_|>\e0-x5O kAAy}?=HQ+x ;D#ڗ]Zh6OaSKef,u9)D6FI1wYG13gW2O cŮ=* Ėr|u"P+{KB 1@bqs`1nY/d9q9yñ֙98˞ţOYip%FS3FNo2-s`y,xR&0$Y&4ܪ4&f,)@hs\]!lc %΃n|=O$9N}$9nKr~*yGqIM%h<|NzݠDѢp-G+( ZYO-oo@QT4? SǶM &,u`u:h * t!HU2$DqwxN!4HFIa7v ]_XQa r;1gIcC A;bbT)QdhIMHi4cFtQw 0>\x v)$uLtVp-4 *. Mi)#EDʴDJMvJ4edBw[f,_2k+hJ;Kxy9HR.A1P$gQWiwUm=koxfEBFeg`b!H1(kP0{=Bp눖^Zxτ"!l"q}Ȳ Cz!1lN!"lYDr5*jӟ 4(3zo(aÚeA(Ǽ@"S[0k45^VFaP V&1qs!EqoR~ZcZ")r$~}a\r&I'_ãFu>j\ &ϟ4 E޽8:^#|IS,UF%~)Ggl}9G?),gj~|~Voǰ4o9:=qIJ2D" :6Cytv5Ӵ[!3:`afz=QF&a$RL»4@zUaoIJR5:a/jo+3*q ӈ~{6@ҸuWʫWHmz+^!izwu Ի zD]\@d~a.ƞPk"`?P6 &{j _IHڹq׍ޕӍbjvSK Q&ӁzezU G9Q?. 8< O9 u}:6{>5}ng μWgwa.W_[>*1`}-jnweϛ;LWt7\f|c?bLSwաa.%;dG',,A`p$jCƹE$ Xҝ,Zx`)k.V9mh7rAY"Xm_C--"(7f>aki$3 $]xЛjl7`= H'R'Ah&q@3>)s:&R7O9^_.'7RK/'9oxyC9Tqp֧}.NLĚo RP@FrCf׳^R-N azPSޟ0NH'h]jv,W'aՒVӊ\tܙ)UL#wa!bnrC&zAAtD{e‡w09Zĕ?WmT̿JX(q6, ?C .*a8Q #V)?J;+$9*#I&p#M:/o4 jZ%bR&0#aO\FdȲ7Kl;f\*偣m>N>gh je Xh4oKbSVBёv޴S6*f}kZIU;&rav޴P&Fv"aӤd9ӈrjuff)fgt'7I|HD!"]' Ws폕^(4o%(IQȣ$GMy[/*J.eXIJ]}[f1¦^PiP+cQnf?'j>?D>,uB߫蟕t8Ty1@ayH{ Doz$4<_xE9Uq5@S/ݯL.h$glst{@z!&5lotmJjG|r|u ([\;WHuV,Z]iLR`1nVQlW9<_ e.9zDb W/ ǞNk!X!GJ\i`&\UX@ؿi}%d4$N= i$)jd& r_*"9J_S &m}V" hHW 0ur-$tf @l4ג%v\]U+u =T:PrJi*NPqjNVsy^B.E d pt- MN`Q}$'jjk[&TM' H7Kd|zjIv&|Λ^ F5N(ϐODH Ye4u mٻF$W3E@?biB݄)R!gEJ*^b"-@"ȸ*22r.1Ǎ>dN*PϸrH;]_OfЎ)07o8 CIqJR),& s0@" AčZ%c GH;)N`<$ ƶ F칛C6;MF3EԀkw\i)&F8PQxyJ핊IVW7@iu} G[Cنʁ*Ԁ q1 Q4WE!JTF \rbB)ɎxH2)PCk0O"IPԷ$XyO]v@rRK@%QQ҃ R(`!0F҄V^ہZYymSP&yY;p Os6)fhH .yOaSpM h2vmp-$K,MZ:lMU-C[*@e5qj޴εMC8Au s  Չmy1J4Cr` nw1p !"  }ą@wI2눍'6:hlQ-ZymlF]%;BqF-g2H|˙V.1 %yV/׭aџ,K%,ǧ>!4F̈́Ͽg ӧLk9/vˏ~Gnx  ށ)h rTHSt$ҭ2@HNHL}!+-F7+}>aOCP}zh>ŋZ@CxQv=YP!8r@t]fPxZ7qǛs/]Do#)K@Qr_q1י$g^'Tˎou=5K09Bʞk'8ཥNMЏe$&a)u>DƣP:%jl"M0TwbHƆ" l>y9SL1&\4]x B:t;0<˼(*:Kmtֱ4pJo :Dqt^ !At5mHW;:vtFZ o(0zqkwT>DtSKi9{o+ԮGqپj24e We-РE 5gS 4wښJ)}'$FE29 fk(8 NDDmJc(NF0V}dDĵ6QN L9 -] i2U 8Fh u.) b$WJ)BV.kcxA'N~9LJ8-UDe;IHcu ̄d:QZԘ, B>eC(>[\=? u028Mt:r<='!+aڀAwYO#XYfT2Քg+dZ^:{)i4_̾n,BR/b503$Rg2Q[bHлdFs9G_ҢNѧ{P_4FӋc% hox\˚G^5+\4\kK ȁ1.rAJl?_`)N cvBV=p,rgZ zs2ٞǡi ?FUuo/e+*//FA2Vv߂`e©:HNxϨtNQh'Azx*vզk-Sqj̣vTf ᶚt[{7g ߅Y{ AQ$ 'uOӳv^#3X|W7H铙d02408w(| w?vpHxg(UkpvўG  )-B58o!EZ#F`Q؞yOY6D,,@)0>2ҳdN)A.`>(F r#ɲi&[,$`q ~ uɳ֔opđEɁ |Ο{5+MY[y1ɷhs`:{\w  S<#n#w6cQKޮǮP%A2tMs]8_X='ʲ(ŝ4)&q&j0A=3)DO?Pz|҉8.Ǜ*ϋÇ[ʐ~U?o> qxem?P P&—-@g ^@~*2<'><'wz/d5QRY#)BI(ɔ&}d)9HI YdxS ~e d-+9_[g5[Kx_&7Y3N2e0FoKgj9֜@-d6OUVȇ+J/ٿLGu_)@ SZ8Bt{їRyrI L"%TTwDh ?Ow/`ZxCuh<.mwmhVLh#%IkGҀ-ZC`iə_i3i!G<6Bf<2k/ ;KglyuƑA{-)q$jctrm4֛a[,ɚCȑ`Be5bLDh -ג+ ;xi Ŕy(aA _ ;oj%o SBߗ0; _QQI)ox_͡uW66.K)hM~\C=Mcn |dMM>9+7f 鮯 5˚X EO`F k^9/ߔu鑠K2B )Ԕfr뚓gZ$H֫&S\*05 o('FZf-k:2{y&&J5Y /FF ^ha\y> %7Dh~k6w֓a4M>p;r`SmCރV3L&+k5GF|2->Hs! @R[:h%^4&-Lr7$Yt꺑lxck*;Pؑ&D&phjQj ?It3+84N,y}$|R )i^IT_ziAJ\~\>"n()Ɛ ,r(*vqNw rf|/xqNq"멝Z/aM'hp3t< T"%#C$ ܃q)F &V"]@S_{/F㫀v,ڶI%P%mW=L_Ҕ C< H%|Ur4ĔeOCq&|㘊|G S"O lչC(|W]0oiXs[vJT* 75-Us ;᷆+B5m܎)!:OhKv]r~Uh3_;H YSؤe'mR&&o߲}ɺ@V. u&۰Է6BC IbMiJ(xȵˮw>-<4M@D1ad䵐:[cfKjC&,꼑J%Th'F)$#^E#OI (s8۰RYŬda]`Bpͮj|ŮB V>Ug1zg}#:z}#@)y+ڇ:rSי!h rekg+4rZV &~oke?W_kM?ͮ9.goeOy%-k E8100Cj=8 Z3΁V8ڒhR`Yya@KҮ-8h"w\Н RvUR^*euYlvt=RVG|z*ycp_d`ހsv|}OOK+xYn\Nwzd8Ʌ _OF"0^h{ahrɰc7*t~6zXIVhIoA6D+a^^jQ>oƛgrąѲaV_hgY> OgϠ0k {B`Ѥ/Ӱ>vOKfe狆b9-[}_2T~E|l/cZ{ܸ_f7aO)FK`L 6"/Yz0ܚT wT+]b1Bz'u@oꡍ\SNk+}k1q'oZAeIaGs$':=uBg:?k>SM@k ɨz_gW+NBtY)=WO'vG ðV}$t7:G߻')Dz~^˄~vZW?_ԵЧ-φ5-EW5S'=FVu@1`#TFg^xU|?KRzzA;X8ۓ Ϸ( G2wۅg2~?fB`p ꓙ^W .H$LJfPٜ<6)CS=4C2 2բW˯]K~vjk4ry4?|b?- VR}X~.~xqQ^ӯ ].0k37 64dHI98%{sPqxM]9s2c{ ˬF{zWh*DufMwwMAd܅[.JUMlOԝN@,TFDPmO6tyb(PI;|:]&2Q<&qisQƘAA^*4 _/Q qⰒ?g$lZ0?\ 3D,Ll1bZ0,S'H,Ni{O9YKҨcAzi ,˜X$9ԒNʼns }Xi=hB1e8ZLXC1{hZ 5C4TJjaQڃE]G"եLdCLW)Q9l5$Ajδ@'j0aho 0`P$@T fgҁ`&VҁgLZHn-#B96A>|LQ8ӐgoXf:}y1iX':?x9d}p2xM`<壋V1xN? ~^ax#C/-ނUl8Ṳ7Oݸ0 /uxE1q?D?h6 '}.|Ԟ*GTΖW|*S^uS8jѺbPDuR(c[Nu[nUhWI:yw,o`R1(:mnqz-E[hNEP]ڶnJubeAImuko8Q[hAV|*SRt~|˺i;aR1(:mnlNUs[rGZ*4+W$ҢeM0떊AImuk81Ѣ֭ U4E$?U߶nvb-BnbVEJQ]W-2@CrM);}i$"QM-"FbDpݒ;ZԺU!_ŤQzMH܉i"6JLۋH6+ 4+W uJ\rXF:r/$RBh8%NKTd#kȩ6a o)LE"{nErD\FKFrb,D fNa*E՝7=&eՃ7.|Wm|ydVM//$djs.#`!̔P6gp>PAr$`Ia*9yn3'@U=[ o+eez/ zfRzÜqZ$XDN94%@ )ͥ ʈ3NLRR;,B$fb+AvHzaT9l!h*8aԃ[a6F{EI!90pErid"D? |['L{-$ 467h$M4L#|!Ba֜"͑Y!)422B&@Bh@&kq}`1enU0 HR(k^ 54VM*ԿL|ÝyLO"~0!K0jUҞ"E](.=|$0ź6P#}:̦̇4AI}~1Wݧd5؝(Biw[wfֻWOCܡDQKG;C֫!0@C+s3.6>$˱ˏv(A\_h7!PsDlfY}(Y#e,OKp(Ǘ  d]j{w W6 B`Qq^wV5/Ц њ@~̕b`H-zgį7QB$ލO??9Ϧ @oLx5ݛ"ߠL)j8# 'bŬXQ TRs?!S#4^"ȋ\e LTYg3cJ+ЂҐ7ιp5n1 *m )X_楺\kB _׏_D?W>޿ !8f7x+F|Fo`^|yB/߹3cߙnɶdƿ!!ӷ_C#y0_a5/ ֨"_hT|AT,7˖^2衬w!衤Ԋy_K-1ޗ lj1%q`\%Ϙc<e,1FQ$0hFY 繩qqgѳiT x.V65XA] $ҝYN0=GQujE1鱸Z"D;ocYB0=/%,x5f,#H| ~w _}>0}f!pì<8Wn~^&Y}v@v,`׋16cap0UalFrn0!Gϴ7!6ZQ {C 1ĭ̐F n<.c=sŠ@-9Kx&8G ShV#B=7B>Yoߦ肨K(܊+r59A0hKigzg޻D<*6`$A}v ֮Gu5*—y3 坁X◇.|{%GOhkyj8o6|9xsp;/݋oa~7wdUD }5_ޖizOƃriѠkvo[ 4UV1bR'TUjp쐨g˽<4^wu?nH_%{,pY/n# 6(HNȖ4j}bwK{֙,VŪ^˺wɥJ7Dy;#4'KZc$HqΦV}}Pcq"KLWFP"yuISHS&0?7 z1BIp%+:䃛=i9uW(-"8s"s-ASBq3-ٌܾK /K2O67L߅y3: )+0Zy]~s Pht<-_@6|O.܏NN:+OsU RX \WQѥJ;{&a@Dc5;wu}Tl~xkȦ|"~m* p1^r*sfǓk\22N E Hߡ.m%v0:ʽ!s v9h/- $N5t'g}0J{E.z'=8JL{R9+ܒ{;)J` [ mZK^d! 8Seܣ<|1c^_+BH_Ta$AZO쓘zb˩!N!ʹA @˕S\B&,\"P}M&TqG]h\!"2pК2a(d#r#3DfّRbH#Qhe{0m=Ύ,&jシ"8{gozL3#>-_RA*=+8~^4$nUC(~'=ς-r4WWSh5XgoڌQ3 Lesx me5ͷ3A'iM ‘t߭pmў9.^P@l ʸ-ڌ?}V?AjJBrAq-J+/Vus›'[\V0AډV(RLaVy{ym9g\/?m+(PVP?VP^7=?ކBt J.Aӻ'xc?OP;qF.0c `f$sD+C9ːHK AYD xN=$՘(=gQaqEğ <Г} r rfB_N9Bv%cN"WùH^ۇX )j~27Dr CNF?yW:,P ~+!"q+.꬗ )pkWU?/La(p FrCnZ 着mli"Rø@D>Dż!,0R-!)ta 5O \=ZGd;zF#?ݹ&!D+&c?v.r 59֢Wؚ KoԞk9@8~P3*D[RԈq~^& !"aM mΙ$H.ɑu\) j5kBXCL6puWcT2 ܷ/*/igZ&EN "!FoQ@̉ epl |s49zK0B6S-E( rs8(8e.Ԣd1ܸ*)YCuK$ſ/ʕ &$G?lF-enf:fW)\"V;_A=ݧ`giWO^QM?`~ ']y*W1,8~Lcp8Aq.nu93zh&)g 2y8k\-YA/)ލ^kpgf-Oqf <p 2@āނ(k_̱Ldg%_3  T)#^Nˉjo;Cz-W:(9?46XlOz@/F?cz  -0DPX0\am`9D%2h0f ɶzo-hxD!mutO8[f"Kr*n߇MEzsnIq_&x:|.o<>\J'ކ=>ϞgoϘf{}Xu  Ga5F*SC}KSOkM䃿=EݥB66mPqE8r'+C8:R a0e?k1c*x)Z{+ĥ[݉>LEpXLGdLdDH|1o輌B8RcS}AN^ñ! ]Vo %owb3#p<~DX5!ZIXC}lY0,~sds')9GW#JW3 ?ˍuQ~.w>uCŹoxs)&j`")npW[\ 4^xݩ}o9LqIBoo|}I@6DwGlM({74 h4p`yf7u[qesz%jt#!KZvRnv*, ^KH(?H_pq6"YyIB@(,$o0Ԧʾa k]Zvp0/ #(x*78H.}|,繲}/iV[t{q2B)-(H В,7#W5 XA։?vǧ+╟~9*2s1ᬰxiaTg}.3KQDh-R%KOǂ{KMme(.'Եg`5X9t|>}6k ZqqT'#4nPpN*NNuluYg{"6#Vm%$߶"Yd{'#/C抒מ98)]|)FO $~Ee,"ђ; /V L* '}SXyId6PM^9`wB .ySq˧n2ys5y2>_!fM<>9Ta;|/1_αY G7->mVr{$<\oض?C;a_w(OҐ\EStWk$iۺ)$ FuRcN23ި&Һ5!_&ȳ8oY7=ʃIu;a˴ڶuKiݚА\EStJQكEJ~x߉@l>och^ɤw ) ȳ;>NfCIX k/\\: 3&)eeXO򔸉B)j]frS~<i{UymIL$AC@qJڪ9R./䃬dn 2PxgILq& 0,۾bM0 UՈYڹ4VRkc!n铲 bűaK`(ՇaJg^?ŝՁCP[ۥ>.kv\#L#XK{/>N uuGq@[Iie:B+H}[}+ϯWy5Њ~pS/c?Eed.nY>B4|ݷyn'oٟ2GKS[N;,)~("NQě=rtksbn}n8ݒT4ζCW񮒎inӎ/׵I*ܺ!%um0#nmDzsvh,p1MHLaT$S~sB%-~8=iCV! *5b,σ{^̻S- x 7hP:LJӻq/?f/%s{Q@r.r139avq0QP.1F3"nAT33(8AHq7)]Yw4ezԉn& a~ʟݺJM@M4/tz< Bro?>O( d۽4{{`9M)BKmaӞ9.[P,k+G D޶>K)#e9Gz=0Q2?kH*aC>U[*$CE$Nl Y%G!->ouN 2gT3h_;ĸdZ(ᨊcXbLiZ#.? !,id c,n9%+)c軏߲ow?_&٬s_!! j V]_ctOi!SεL1&3Za6m%V(g^FVh3Ty.ۆԀ^wi2Sv$j\9$ W,T$cҀ}A@Q>dH*\Ձ莕l>\CS}EYLj/En+5${6N8-pAbIAwC<o8).@ ?蹼EIDx 2O7p+I HtEC$1t9"nנRm=u}ehrm{`Q-1*f5\0낷xcF'P ' vxm^+eL*Kd;?6sd)Z~[CmL4Hh p@n1%0B=|HӺ;'0Pl W$r)M8mFۯ)QB9(Tm[B9&.MoȞʔ=E+vL(E&y;q P%jBtySPPTo'J:9Km:-PqNkqAuZqRFojAK6ADR#GљPJk*5p JzJSNГ2:S5R,}>]ԡh>xt[Z= LC `0H**WͻݮƆ1,j$d̴ Baa:[Y+ x5M- +&AqͲY2s[' R3bꩱv˘Ls T56$:=gesuTѷ(W|/tφ~,.t[+Jsv G$DP8P;W>ە&_VʸmkΚC)Vpm#pOWj\LʿzVz"=c.tAQvcm Jzmk׶,[)>b N1(N`Ǡʸσ)uH1x}JY12g2c0 ؑس:фҶj|ƫd[bsz|2/\ub/(yJ9cmyu88(VYݝkzGV)CwĿyw Z\on|JZ)vx>SJsz}I@icyIni`Æ?;ȹQ,K݈B)xp4%޳UHBcdǀ$AƵ4Ns0}= 4&gql$+͓71Vc0j!!8 *%RHg鞎 K"ň9m AH~DQ$t1:PT,t`D(8*u`GeʈX #`Iz,) g؂:=0X@)MB8xaJю3+EIU w yn;}C5'wjBI>JCrEg(Xk";L_+O V ߄ys6 FINF -nܚ;(2&"k#F^P>+ i!m?"9 mѷK{XFt@fxth_3MWӟ[Y^*\?F^5*|_N󒭁E +ܯ`}@e ,H֔f[o<Oz&pvT{ j^G |35zsa昗t;u&P-͙x?iLZݞ 9B5=TAhI=n BBUʂMw: 9c=Nv"Vm5_5Xυ]<,FOEŠ UՊ+9elw~~{{`BRv΍w~]}u:w3A^2 l[;9sq;qrM3ӿ5K_#6 Ǻ ː~oa i}|Ļ^ag8 _&3f|eUtsN"89ӵ \|nRѱVux©Lh$ɟAӿkBX2ƼIRf5Q}l&T-GOxY .x:RjGZ8=Tkȱj+鮻A|%zl8h|}Tň{w|2I7l'ӔvE bb,Nm?gÑm@/+O.]LjUҳ)-z6lz;^vS/>O^Q /޽脁Lmg2o`cRMTc*LeE] s&QI_LE&uu*.8Hv  ʼQBjm%fB[# j9&o4 @pQJmD$J8IbhW|Pʀg|}#o7$yL"(vȈ)a|u㟩w̴R d%ylq1ͮ7kh^cD7 > oj`'&[T%Uliބq8[XPi{jXցk N%jF#%30_U¥•{>3TƳ85O8cm4)hBcY \eV7`\^ ;KiJ}N1tF.Ϯbag^}zgxfb:raW͇6 `5:Ḟ\f.5jRSY:!c916/&ks_{\Ow@O͟xa*߻Y/\Ur8̑}SYW= (ߨ&^19)hrz`L>!7+6,_1PL^WOw/)A_Wvj-'H`"["`Q8](…@a ZX8Я6'Rs*Ev{6VU: ޹r~n*OwlB)*qPLQ3TQs[.SԼq,?edjon/0#Wmw:)tٶŠ ۝˿}cAnNܹTKps J2\b#SjhFQi!n\ٜ̆EETߓ+.R4Xm Tn˥198%NJ 9N#EᙰZZ 62xf_z (E3]aQ?70W]T9@=Lφߎ>4b;Qaq(TKz|9 NXwv z&k5*.y4:S [qĂ" 1R y _ ji-VzƘ DZS Dܴ10m31'fPR sBB"OMV[cAL ߠiV\>tC9B x4Jܮ+@_`NlCaNLE/ EEҽҶ nl{wi;xMX2EK>iaGԢGbOE`Ū & y蹵V`}O]Tst )}B h9kW >Ez$Z|\WPM}A G13rn~5鯴EuA:x7Rͫ$WI4h^UEZFI#`x= ރw)pמ̡UpoG!(m&|,YpAGvhXmEW'EB8S~R,][s7+,df*ؕLv_v mŔĐlg*}& )vUbM4p.ͫ>a7ˏ]ۻ71*Һ);xH캇x˱\5ĸ"yq+ !&%!RΫR[+ECcW<4]j+!Z>Οho R"ULԫYw;[Ŕn&ZO4b2>lxw?Z*~Bnsn&+oxGsBL Ň0*mHcޡY,ARI`e1hY&iʤRCVb %dst`=xnMܒV"2D\!לRE[qZq&-\ͭ_ sYtqkn/L{ݖ2I_]_7.;_bB. ޢdm:dqᔓ${qmd\bεZٴf -.>l%<|ukλRݩ_m':+]w7/{Ui3 pDSzMhtY}ts5Ws~th'FsE2;>^()yM&y:knݛ) ;}K,&Z8 q`ޠB=+[kgzymb͍4y aQƘ[rSVl1.YcP ;@ cɒ'h.EqSBJ!y{;$\ k+b&jvGMRڅyvL;Sk'Bni&?_&.ܱ3j ܪZ) ":rI)<їN P$C:~M 'sEԝ-t"N}i4eEH縗e*R]{;ZBq_= ᜝K:cqJ%9.DK*S9nd;$'c^_CNvƒU5;kh%#I@ܽQ47_e(z^XiW$3^jiFyai%zJVh |J ֵntA:BL+)R53$Wg>M%{{HO,HPTc4<CM g9T9R<+p9b̮_s1%NIz0 j:TheEZI{4cw1r_ZdnEZHz[G-d+Yzu@E嚎?OR:5vfLhK#KzgKIBExsݻ \*QJ_;R+PgVc !\S P8A)xJ5,x9#fKJⷽ!0X!S˝W 7v]bf]+  pS>c1qW2MskfFʸT #AkK'rN@5ɶô=VE|5|coDw|#L%iEa.a4'OſY\<8J\ͺEcޯ&^:@&\/}\<Y] 2BCP"?r FqnI@Y1 3e8L4>+@h0,'>{BkL$w^7 H/dQ1lWK*JȞ"ZR,F$qt"rȒLR`,Q"QK)T9b?)U٢*Pʝ23RQ=ޙX!V` *wVo1R#ָ]S% PNZPS,kHq4SQD!ýQE:q)4یV FPq`NZ ջT'! VjP8~;(o(P( wE*t% BLΖH9)Cْ1JWmI& 7(klIp8;iA-T4nU"S%[RFun=NTE.(ʂ;y&3p7w|} vְ\{RdO:X4b CБ-:k˦׬5pzk6kٻ5C*#-l|5yMj`ģiG6k:F2kr#4ӂ娲T$gf5q걌O4kySե]=XxUÉ a 쭛Yc#]yd_p:?x=E=^fZ4AA5 \fw?y9P&)VuS.{kL)OG\~j4MR&_Ah\l^*<!$b ^X T._եaAH-k dJ V.k. _}X⥿K7|QlT?Robb»ԕ J055d;1c2(ʠ VҮޛv}0"): 3ͼ:O\|p_E=Sp-jQJAɒQdzbJF iY=VҗQTK:HSYE?]jPP P_a ]ZY9(f MW@K9ynx() #Q~N ۱Z`ȉQ46QmmÄ*Z\#ΩsT̅<`fȘ}hfS8m0|v<J*aGARg%n)I.FYd9FU;I5Q+hþz0mZGl ; Ƀ7ōPDR/ѝ+ hjw;y*r)I(B$:[z72D"e2`fTC#ABv2 RLVr*VۼGZ <0C <2!'85Lzc\K+ri1̶ԜnY˒ce IPEG?3 =fC-%&א4+\6%v j}vm]Ȃ̛x:|-N (h,$J;(=/$%O IT9](&N~JZ%IkUOuJd\ )NIKj#ԧ<',BIO)ATPfMbg$M2g՜Y͚b5㗂Fg=5}9L0Ə{siwfzOT7h.|>wdk(cB>qt;X6,нآ;Ӗfy4o O8e1_u\ ˹S䒈Daˡ0F4]iF~&Sg!{ˑ vvo{A% K ~> AOoE< NҞw<3EK}s33RӪcM,~ ʥm:uc~wsCXP!;963rxGp. _v54!ŭţ?6I`||4_%Y#NΊd7pdS_So0}Cl[{ !Jӡ:<ۡ6z ڨpRG3G8r ݽ=XcCޖ3~z%Ńڋ]ۜ6}/X@iWr9P+f~ XBܙxqot@_|+ZN֣>դ؇^韊φ+~D1# "j|J>E_quË`(wOX}E?a~ >y=ίVk7v.\߹MRwۻ寊pV(Hޒ9Ш/E?;H~r{3 \8 v!}.0 ;$r"T,Ǖ@_fQP\uc[-_^TGeRKLE^쫒7YavZotns#f;7szEhĶO!ry~b^Py~ځ;]ђ׎h8rhMG'BkJ[TH/~9 -'lums}tù ~qH5wHPB(.9yGQ20RSd}wZ{lgiO x pCR xirl=;vYP\š>r Rɤ#m gp\kUM)Of禙  -PyL/fVUG6lv1nb:wzs%@ԪTPcTgJ8BzU.1` Ɠ͗uZuDqIIMnrےI6]CxZM^HF.71ͥq;@לq-HL0EOF|5A<2>Y8]ߖ_E٨Q7NnΦ!ݿb3>ɘ^߅'%8)!I 9N!Ǣp==gdr2)͈,D&Ɍ7*a<*KBϻ&P?f_jwx4>BŞߒuFr@>9Rعg 4RwO*aT\!2>2cMۂeeAEƥ$cEՄ= 0As87Y隠 Ph'8(u-) (-& ud\F@cr'PְVq.'-Xv(eNDd\=XzU]dp#ʾzC91 5H fy!J/D5IA$˻! EP"_BSz4j~7 Wr7o:/ԏߓűГ}>6+@9wZ<=џ! 􅆕V22mhpo}~zik|9Ey|aሂB~=; } o;gfbzsZ^׍aWZ!1ARӚhtF% .x: 0L,Yɟca JAJʡOd5ϑ Ge |\lC項^Һsȭ$1kc\ 걏" FLT;eN`EP 2H)q,'Gn~+4f'y݇یT }/.Ǧ)S\DmVl tHuI0 Q"FFš|٦6.|#!VbȏͶJ0w]I|YS)4n_^Q/<ivuBzYڬRiO ,>Ƥ F&GaE潉Nq|uV$Q`720vR!N %QLݞڕ?.öu=wb'gLꋫcccƞ2o'BV>6vf4LY/PF}\;,YUϬ7y`2c؃U ?pqw\ԃ]qgÇPwI [1vw;TK*ڳ|)ZwrglM 6TO% ;1U=hRC21OB#ׁ9 Pʲ[zѪ!v.X0w 1hz"|^$daQx^{Σ)10A7 m_-4JE6s +z ɝN?=c%r[G2EVnP rU㱲rߞ,0F6UN*)˄y: I>ĖO)Vkvf֛ QaIZh\z[99U22h6ײ 3 ؟g3]-)U,lW: Ahy.6t!f`$W§X.^['-d1'a#&/ǖ-VgWeKT[Τ /%M.~;G yV!mM&>s ];?^lι. '*Nj uG֜;P}=4GY pp ?#4yRXԑ\~jQ._ԉz/.V_~2bꏾ[4;';qNr3{,1ZYmԒ YGÛcSzqOhz!LC#'XLM3/./@}~􃛝[ c潏nz~O=*= ggDٸνrOmd< ؝>Vk靳q0^@ 9Q{ ln#wn5^wPs 1`1c4 ӖC| ! zגN Nw1pnKұtQEN?i-'~H' iE_nVoVV;S׉8}p z|./h+`u*|-L]'$"F ۓE/ ˠ?/̝3rkH[J.NQacsp e7]$^_Dus3}qNhXÔ)[nd##v15z)n:ߪC~qӯǬV} ݴР/? $b>TW+;9Aw,ҁFk4i咤iH&ẋ37Fח>v/4j1z/BmKl^0@_G!gͿCԧXrzcjf,7q0vԌus6,~.MVVN{āoCM[z ]_֠VS[S;ahxl=\qվg!O =ZBk^5F9r cHOˡ.*qlԨכirrsF ˱&1"'EO蟴E@^35 ʊ( 1h]3S01pj{m6SJBͻEѷחͱ48]sCR1%G/4:8sϖvd#}Kq*uEc)WYH&rG(UqCY͵:YH2W) ,{z d$You{~]io>J{CC +ukJ{߇l[~pn#2ZrR}5f$ [YZ*J6XbMse9g"uee%i]ר˖ȂWnb 1>X1+{տWK<:2:Pa, 4imF!Wg@~oL r6#6^VGw{;/=W-k`:)(b)4l5_~wU pVc !-Z 1ȶXG%ICZ Ã):-[_FSx\~̦omPVPN1.`r?m"!vUHS=0 R')AVX{gAg` T{ǔhJ9ݏ\UZM$e-E=j|VdyPbtLs rxEr)F#4 [wW؉d}#˜`,{=~Ni(7[ 混GWa T-iыʞØnWsnf鮭YB}lt6?8t^%:+Z4x4.?MSw.k}qЋA^??ӷg'NZ5d_XU}bj S} m#ܻ~J>?D[f y>Lo_A{k;AvnYxY9N?oT UM1~+RYi?9Cf<'ٻY/-6ok>ycl'g/ck\|ݳ&P⍩\@ol6MNiɽ+5L?hN>Z}  n_ZٳW0o,֎[W^\&Í[Of@[ҿ7áؼ?7=1q4xm$X=,iXS >z=[n(uʻGa:ftyr?Y PC ^beRq. * N(hPfS,q![ᄢRDuOG/SQ4<=RK9ߓɂөO KMD1q nRbR5H$N0M" jwU\,%fE}%AX75X%&B#ʋgZ>?ˇYRl)WI^(ɹ9`e| jԚr9XcY\(TrN*[xDΑ:b F΄G)*!gl9 KT "(^3U<@q4+RL`LA rhƹ?^ NQE`.(4M(LN42,ؿ/{)dD;6n K>fQp"|d ;sۢphDLpJ?.q)7E@בMY3ȨEkXYHя*CBBb!6x!CH`N)րX;N4EÆL:QI` !kL3){IrjTE(+ycO|A4{ikeSB&8Ye(\PjG U|cQkA ;(/U@7H%opFRLٷOؤ'k:PM5qR$@]Z o80\)c rڧs=3 t3^q&7{WyB.-,#SLK9*L+YUhn w;{t-mW<-Iqܽ&#{]R|6-  bE/ V pI֪*ҩ.`8\aAlsO ԇ\^[Ɗ9#"!4i$%6hMY)YnO$ZyOGm#&R;0=f#(%`@ k$Nq I<kp*Od蹗s/y;Ke8ҦŔ&N&JE /Xg51aʙRڜ70׈r\VW Т ZuJy?થwŭ.&8bԉ(dJSṭ5}kuZ欮"mjuU4* yF' +DR̪e"^FϽ{=v9%1Nh XNX( ƬJ$UI+;xBS*Њ.Q/HGz=\4$ 㽮Z2_\q~ՂfcJ3.w&0ȆF9lPt |p{і9 sDSĄpW%–tx*5[axqPŵ m_t ԝ! +g&AEqeZgf FXi$5Z)B\ e IMfM45 Ii3cYEۏPWuG1i:Dy/IWeHQŗR(O* #C 1a`--ǀ DhZ"td*!+||9R҃#JRB%ďo ~` S!KF✖?Ijc RJf&BwU\ 3}>\dT Hܨ* j{(cWz~\};6;V 9LQL9 7ٸڙ"U^gӌa:xAS2wj^v:pQ;k.Z;+rTehPv9G {ܧڛbl~@۩u|`Y&_ n@)8A3 Ν>Z`R}aoZg@ 8Nf c xک9>ȸd32R9}7{mPo[7l5ra%En0_;^*;vk-{wWvoIiZ +Qsѣy[-ܨGG7];`> $i~%=xA%=3\v3tEP7b ;̱2Kْ"L)lur cؚ':dZd@Hl5~>k؏Q.go__V6@VOBv+9w}Phr<{QxAw¯F)ǞNV574 fby ]yeZuL0qIib21ረ /Ò@_-:]DS:!io?Н`r)ߙ/ga;ˋ=WZW쏷`ăA]x@(\) lWfL H?lȮHO$fM95ţ0azinM}MH=j8TF{{lzR8Xe"cl G6!I(krRM"yleeN}dk9f+Tkxd|)ˇֻ:q%"8bMcp_h4*g>Z9dfo cE2*H" 8GErZ2R+XˆZ2nU°)bDi2JiI V)] $w0o(I۩U Qc>ANB"rk 412UreTK[iYLEQ,Wj=i+P"FWcn3RPcD n$c2aFє!jH"j)(j eCisȾEMB!D|=Dl:s&Bh3RHk kVFԪR|QtA:bxA[6cKlwyS;\[є_U!(ElohX>D3wFxpQ)ҹtemD|b8ޕ7ERL"Gp5[J,QAy`Ok_8F1r{eȖ#ê2qv<}+CygkW?tRt@B\YzT_̿1XuTƮ royM zhHM GENcV5VF$?MOn)w.$Nй2[Hn(b{v't>z}9,J#K½D|098퍯6B*+4a!rxCd~딸Ef<_ܗ|s|뼸mNۋx0FO7ĩX*bM`Yf"4|~hMwTg9wkʗlBrO6`8wn XpMMqAX-9Yw֋݋Ǽ愈M$4RE 'a8P<ʼnXL(ͥɢPR6;',$X}a:'qJ n\%T,M\%rAɔT-%¾ >XhMvЇZp;1"UtvJkg4eltD 50 ΦA}UM.Bp= # aSgٙd]ըD)BLlԇ+Sr6V8wLS<F[탔z' }|+@@Z-1%ߪuN_xst*~m;FçG/<>0঒Zg5|Q:ӊ۵aګͿbG;`N?mu]tUruf|ngIOGl"33 pϒbz%w 4, \F׮Y4ӽiM7J6~ Нh$8˞SE="m^;ٻ8rW x)y  </MXHekֵ/rleaI3?Vn*{,<4aC@BP"XYׁD^$YDٶ,rOʿAFvs)s@tjz llN 2`;TsIrpMo<ٚ/ ێ D)U~g 貢ߋ^ݙ1_38aB4>!G`˛O~IdpLAN ]z @t673qɋuo>{ aTk{"U\!Iz. 9:$&`M0j 7k秷N^o -oD9yvP]#j{AyE,u^M'M!3hUnP^@)ipCeǣLU^U 2f$߿x~Iq*H9BBj|KTZBt?A?,=DBĨV$buF4: ǁ(1OvD2@kuwq';kfHޒ*c;os, E.2@v4(`E<#@vH~ 7:ƨ$gY $n26N~5@?{"L?1jG%Fr2CKtMǒr+o=wydFHs B5BZťs` Tk !!SٜSOk&4=`\`3"l4Hb0.IߥԢ4143qƎ)C!CpeӥqP넜3M3M(9TV'Z6ϜC*Cj|(00J е$t%69 Z6u!sqrĦdI ':EeM0H:%fk @,ya{;:T!x*ߒy ;cȶmKXϿ?mN\TgKAizeXh= Ix)xAE˰8kS>U\lSQwj˰8(JwSS&Y"c(YhrHqy!ae#v +~c=TT(8EU!U=7\\Ơ\n;7$JR*v/Vn.(|ys;MA4z#.vX J^XN`]z'rB1E:Wfb FYLLʬFKQ@S]R ~.B$eoFL+}mb6FI]@_uس׏1^'pdh@E t:NJ篁dl)U]hpۿV`Єéx``)3DҪG3Yon "FD|q^|0LǔY\C2"` 1=@{N& .\Rz 2i}s/8p2UvpE2UMh1J e@)&Y k[c=z|BZu?}NoH=o":#N 'B(jO?yaΝCIYU42'%m_K4:ϒLm%`)6yBBFjnؑ;p;dnFpܜˆ{Hć&0`ـx:pDPO}o׸v|{Qr[4,5F4dIxL8 MPJCqc-uefxS5URui HK~ חh>Ѓ•\gQ< idE0 d Ev@JuV(Tez]2(VTn5@%Ss.E9Ob1^4֥T6}t*1xF;z_Ċ8 ~] "jփB>JUVs+@JRCIM&.44`ɣ/~v[X|pN_qzpp'#7 kqcyw]?7)jR . bwzF K]c5"7L~z.}r :b.3vWs3մf^]v! ?;:](.$ ?K,"kxaC$d[(Oܮy|C衸hZ41 ӓ~LSbtAjfT`w?Du9ѳTňILl] ȥ|OP:'Zs*4Ih8Ud`[c2'0wr5_=1є|(\ԛ92Ds)Q+бR ԇ=M/^}rfW9Q@n\'z}8FeE9 T; jƖ@u/48Z4=៞kގ}4+{Jѿ}y[p-!_@$]=>_-}sO⾬6wּ i8fhsl28aэjpJ!ak$zŵNr%who".DÏѢm{}~r+G~?]oruZi8j%?)vͣ͢,uY-Cӏ_Ji4u,QZ$hl oq;;c3FSy<drcҟp' rEl.](k.>yS`梭5?\E[?hq=y~jbF=9Rxf?L\kd[2 \n6iJvE|a1A碭f/Rf?ILdAkJN6CnRDI6(C h/miTJᠤ:ՍOnRyjQNq=mL'*P=|iC@ 6D!!/77hr E@w;drVw۶QE6coAXzZIIDTHʗCCQG4bÙ:UP1^ٌ@ֲA2[YF8 C##!m[M#1(b^1UgfV5*`&Nd4Uf3h4:3HE[J|=c&PBjZc08u,`8X_Cb\B%و[[{641c!96L@OeM u魑ތyG%qː Lȏ@č |lPk&Vff L2`ǙNdYU(FC" QX$KF$9Xh0ZI.UiJW F7֩g'f)K5KYYbRVD^LFzo/`X8CA6ܠ74sH(A/e !hk`6&Y SX萑`v)1D4JM6|(ZRVgE_V#q] =ILFRbbgxYX81$dOQBPA7֩}'o%,K^2Z1p*C^&: IS죕EplT:]P 2C] y!dYsG+(t8570-W$rpBo'X?U0ivj>Nͧ۩t|`F<%Q;"8;.% KK( p 3c✗NWRCB{/xt~n:?礃:)r  .y >`30R/iTRXpf8P589'@^i3g]"X_{瓴Dlٽh^ C\2֫3|Nٔ+swxj-`*w W">RH-P6u_Dn'~kgp2le]VetG ehŋB ܏06p: ? k3hm#8׭8 f?9^w=9zo{Z?{?k:ݭk}k;xnwr񂽷;?kr8&0x˽u.vؚ~:_t9V^uFW Gո52qpo# sfns~&  KtuzN >3S}|@x~a6R0/sdc:L~24N_tBe,[ܼ/rrz s0]5\}3:m=`A'idX_UbWMn0蟛S=xVgߟs2h\Zڟ ֳɇFOc_rj]F_~GC@t]~s5Oaji>`_:| hvdk?r9N2I>bѭimƏعu டlߥM>nȿx?~˺ڿv@>^ŷ}0 ItwFyC hp97ӅzVvi旷vÃ:{?֥םŧwyG~ࢷ34~}gNQҝnq6ЙXk^?  F.W'ɺ9; }:N{]M&;V?h-טi2ip3]eLa\<~;t|" F'#^^ȱ^^ k%e)JR()+FIՙToxf|ql}@$Wo¬̀ȁ EC!8Nj ꜟ÷ (⸙seFfuܜ (x~ks=n醿},ÎJDU?]XlR&Bȉ"!:c1kaN0bp k6'5v k]+Mbs֘Ɯ5欸|'zq)cYHR<E;h,hѶ6))a-AS1XaVh Ñ[er @sQkLv K_q*ҡ ;~H[zASuz[IہqSB<&Ex*"DZYâ) ((D0TN2LrGU;o(ONY%Y}-PNxVϖx*K S 7 (09Y0(o$J!ND^]WՅzu^]ՅuK.0L#,EY9g~wQ͵UL!0@B811HGA 2ry+W5S‚Gk,Xc XƂX\  ,-gi<+.~WeufKepyJ&ţ#:3* Q)ysM9M)N QV+ݎɠ ߦ68C F$` Wj|tHi^[8VKUMmb fOΗV9pÄuV>h&8!T KJU#1gRu8/ک:Qu8[u8{"@Y[G,gH@JFYz-=S,znDЊ"-'UH_**Ìf2F+5 5eP^9u/]R!([dZ )ND[R#7Ch퐡 6aCgs׋jiHãw[_S_r`w;ow^N[?_0?a;~uO[0SRoFft´b9%|;̡fׅy֊yBC$kvh.Dz)# Q𮣞Ӟhn3lC~*V z crԨQ7UYPNk5NbNmnu,fu;7V'\%xŰwfȻH- ,)rIٳyܶhe "յmCeIMБřlycsŘ[4T'8>>ni]AylBy1۝eTLUQaoYw6W}c{5Zw95*OTqJj$=~M9>**Pۘ5u))6#gφ^yWyI; 9_ҞoTM2~nUE?l`.N_Ega6kW !LFf3-ͬW$ P*/KK) jE,)Xj FlEIx~,aN]8(,fZoS&\MmrJP LW"?:0h 8:֊@T2LJJ Ӧ*Aky͘;HUO5hYN+޳>֩"]h<i>n whfn|2e ipr"l|qgf;M,*3vN?e>n/ uSo*%:cӘh)EקZ0v.KzM_5KҼ4k^Zf\Q_:đztMKFtj1JI/Miw5T`Xsѫ*i6N{u$:;mc*3͹ʄc JIe#ӽ@ϕܞ Ʊ1Y{ XǿnzZB{}΄YҲ8enEI>qXfE9X#ݵ-M+SZf9z'``` {Ta%-miiB k2R#t3 bLp&:ч,mK[UYn:nwF'!?\i\z/Q@JD 3%77Rou^w+xmN@ >fj-\U9'{$-,4knY'ڐ|rB/ܺ*qTe`x] ?WcunQWa+GRՙG\P&Wq=:X+LՕ[)Npu:5Vuf>27ioMcH2͋ҧ1O5 n% *=$ۄh%EX@jsƜ>>E),\i䎭``! י^:xwQp攜tF< ٌ./xFbt~1z s<[ϟZMv?nc;o̟JO n4o0l3_T!2g}fHsskz%EsJ"ʹd7mhQ=D1,rY]f9D9#m^q2ϋ;O /qJ8Ӽ ,d-QZkJp:~ qbڝ}0,uc꒏:p/ ]ҿl?J*)? b-)L- uaEALk:271p= js= ı`/1#fL!`ar VW6@Y&[%yL@Tj`j%ޖ0:#}V|rtsq 0Kcct[ƱM$d)ia˫,WK#Ikn^wވ٤|қgݸrZ\Ha4x|Y8uIۿ(V cV#͝K;ZkDKC@NF%e*5Bh7PPVgSl{6?2)=ugs-e)cg3]Xx*32h4:3&L\kw(TP?ՉzWf|e^B軱_;mNJ9jŃn~0Sb6?i%qä޹5q#{rKҪVMkڇ-mRDJwk6P- 9$A0$9JfV=SNt@F;:x|=K"?߽'ܞp{ 'ܻ#%ʪr8N6*{ " $ W,rPVq) y [W{sB!e"]c$Kҁk+/9!c&#b.ȔQy,^-IwnI3OO=$4?ۓnO={w1+QaTZ0,Uo=uB`h)0 `0^ޕ &M)YȅD.p ]E(19*=9@' jcn8=@W>t{Att B0sٟ\-1FSN \2TgѢM Lz@]x%Ln %Q{s ba'*o{VVOn$>BQWp|֮hyvYaȉ(dH&`(uYDЀO}H݊ez%q}=\smϵ=^NVN D1TT.Ƌ ER4GuEuFk|T B0FJ ;J؏UVhaj||8k&,"h3bRbs"m .x'}s[M=Ҿ$ߟi{푶GiiUedb5r>|(+JUMAQL6'hC_O%0-%0#C-Kk[JKDJ3|B\0&F': 76B/$8L,)[5Er#ɂ=ݖqzEm-0=pm=.(sEc O%t[\JrgH_X.ܣv^`nש5/uMd6roΠ}xVwq Fq>3j,mR)V#jYLv>QNC|xpK0ᶇn!͛uܘz6eW0PbaS8 ߴ59eI0B18g|Yw3U?ய%{s,ZS=kf!)Fu3b1dtP%EN&/tdB-n:=H:n?{ۣ e;z)@qa6MŸ+6P<'+Â˞9IEIQ$޵v#/9"<$.`yV}y[n0rFas?dJD;"JNiDh,|)@'׎Jv z5cK0s{1ssU(+t$0iqVSɜ:Epz.M6r/k5h#+gp"[!ս 5#,a}\s]$( h8a8Xr ddFy\~0ɟ-I(  ,C3X!<^.J3pvAk{>Oܱ͗Z¹:z6(-<\;u=pKSؒR@^X %!]<[u׽vEfgmǮ ZmI;p/[@殫^{k ׇ g=@n<5OC+?{ͬg㐒rx2y:;/If3|&typ1\szc7t_;٢A<Ոn OfG'yF?|Ew65pg`0GW4gC}b9QF7vXX8L_vyX/GywxnEd7Is9L%l FDV|E /dwgk>颅:Db6&QB%7v;q=kF|Vw XQǭ9^/W4~N~%.7nk8_BplE_586Y^N?17vgA2yn V?xaʻ]L +"=ӕxfL]cv^܍5PĎV<\tWcR+?wXԈjM5 J>OEC 6ޫBLG}J\Rū[=DIj?˫ra3a;G W#ܸ(Oםߞ "@5z[R tWa Ŕ&@fśTr)hGcT%QmƆ?VnrSNoY!S@mRL8sM8Ѱ0Scg90L&۸I>9G!֥`&PQ yN!{Mtf!d 5/5tr+-nR`/sқH׭3wZ}|ށI79;:PׄR]lb1zNYw'VK}wQ ]H<|ƦY 93@.UVS}<֭&0Keo? ;κ/+>cF.w"Vrgjiv৳_S]t?#e)sL_O&Wm׵ PMw~'$Sxozfݕ7uJ5dXIUa-0.&QDF>h@͵/~dD ~<_ٰWW6Օ {ueʆRj"(M]Id_28'Y9H^BH>Gm|݊*u@sg'2kh_Ϭl?/8uA9>8ޫC{u({e-e/zG gE6*΄HݞZD3E8NaRFG/)k;~Q'㧫ށ6*H$uY:ptTe#%5ɼ*!PWbteSC.T4MmL&DŽBA Fex*s#ѭ !Đ/gDP ,@T)%D͠UCdVeGbAh((qbP-DZMy4u !, ˴x۱tE0F`uN7d j1[F/Y@"K8Gd # tõĹyWTP4t 5M3o!u]Jh4ZfۏC#ԿeLe{g4,ԑI\C]F6^XE7[12Ig"0i,ջJeP%]w ѥ;= md[%O4۠J,CCYmP'ʸ"V22k2Q +BHիbP&QTԮe ;bPvaޕ7uSaBX %ܾq))2, 4gR`%գ[hIZr8W5F%{.<+HE>xȓ,t)Ѥ(&*!2LEf@je1zId"f39f ;IqίE& LiFkdfybܧf-p(̀ >)?Ơ!8tcypcg-ʴ $m:/=^ 1V?< Ȁgm紪wa$R+AlI` DL,$2r-9;bɱdO[el'b'oAc&f2@/!3f%X1V 1D0el% cx ?9h,*oXCY}kV LQ-{?ᗦlHq Zl-KHrMlxF똕2/kfFhr9gK,a^sgIry:z . ;0q}Q*1%nۄwʓ c|uF3B}j GY5G%#tbfA\_m/~z4UkVSܚbgKFȹڵk 84`К¢dD`]fV7cJPGr13 $i@!==8"rOs`TBe# JVRԤ-JjѺ X8z-3E=rD= ʠ@|28.U`w$6F+r =V =9Vyn܃xeqvjj"{QCz4CMƆ[o@'J3yP"0jfS{"Y\vhFFh&2wl_TńSLA.<=4'se9X-4cL12-}:23vbދ_1yUO*tG\zsq~ ŧ{{ݶOS]sr;UbBJ]VHY[ol"S'4};dufr!Pmq8(7oT@@=O-#`OK{:,F-77?J˙qZw \UWCѝsvIiJAjJA)[ؘ)bct:XVmhd4(Kq$Mm8I[^i&z:&)޼د[ƥį|+kEtN˥dԦbde1ڬŢ^Eҍ+U,4].U a+L TLOcA~ :Lx춻9 ;MHKLWS)/14#ei E%]FsZ0רKLz&0 ְ$diBXk7 fBe"s$vJ[}{6(됑2tN}A9&5LY= LV3LHM+IerQ|XX|g"W*ȏ4U4+nNBB1ﱧiˣ;s\x="wX  RlVh=6HoQ[4K1jU4uPֿp#G=nDȟ?ӫ~Lr&}Q\L0A%٧c:.r=9m"/<栱Cj9K>)¼l Lj,HE/=5y<$i@-Rg)]>7ǹ9)P ۧ:SFtf\ԥ@k1(UEJt1O&:HM55*&dW|M)p4') kֺ͘FEo EMKLB>=UcraG\ 6s~=ۯ7sϬ;Ŏ+>KwQ{qr ϒcxw_`'vysg-s!#q_^dIHY @Ѡ*hX|oRkoSϏ\r:$->G\\5IK@}mذK03֒1, 6_h1]CV4d&[6c s:,I*/oc}Z!h\fR\fB.3^;&o1LУ@9 J gflW2Ӣ~ To!OOrMsbI8LR>F c\>4 0cA=Z *d՜t{ujQ^zJ 4]'c֘NRY!-]ZAQKP"@&'th3AqH'9.RZ Ic^ 1q`CL=0F sـZ >cpup!gcC#aز%czmd2e-dG jA&~5ds@&xi#ygݗ-`g+%G̸F Z  ŵް":Ĵp%q: h$儫ZX]]픡蔌y|HG5Shd]Q eR4稴Vz%Q@if>M']0+̦ ܡ9ϸ ̍I;EEWґX6h-+l6y|8Q@{MtѠe&أ-uP7졁̼[n.״sﬖڭ7Ƴ3V]}j(3^.zM *uIp9{.QY2 Jٞhh-9x%C'H ZQ5Z-wRFiQ- >0Z[]˼TK@fjԀyl1: N{?XB>wJ i(su 1^__wx ؇[Nw}+Tv%Qow9&^t3;kR,'V6;xO.'W 0> f= 5r|t}+&G}԰]=^(yn{ur<ݜw]:CŜ?盯ofKᱸ ~7Ç@FۧoeGY>w;ٛϫǻ/}w( 9.ޜ?./~m/ޜ^\嬩Lᨸڨ;UBR%vh n݁w7ķ!J̼I>Ivw}{mO]ۡ_[=(Mw?_kȟoؽ^<|s^/XZϻH[w>e=/i8i&fvV7Y Gw.nl=΄O_=w'S9:רX]'ۜli)cs\iP{DO(9`}hTp4U2hYR*xjmTju g"&C{#(}&0PAib%,Tr>i)D'Z(Qlm n j.2\}Gdж?uđwFZC:zSB8)a#n` >uagi8,A9]PYcチQ8H"CKZV2<~F&o{qͬ8ys1ngzn6j'6C&h:'_R'Wl*Ĭ];ǖ:.ؖR*^O 9{4vEsU:Df,17cR#[]p-"$ɒMR|e>l<6LJk".EܘY$a7[ARi~גcPfśߵRsKaBfȈcYi)k6آNFt*L vAfд9vgb͐GxN᏿~ I<x9VXj͐>GCO/T\m4;`Oڸ\R;,uUeg|,z ˡ] u"?i甭`b$_d~"x4ڥO)~ i4Tb?l3dѰ`b6BT-R D?3`_ҶьgLJ ~?$skK?c;hO MFґ 7M&$g ZܯPAІCtPIl`K+*Tw4G|zEا4&N(NaUBU"mōTNP˽#H77V`&"uR\ t_]V2h rw |<[<,v53d`[?T)jie)rWcb0Bx%䏱py#I鰜F֙ob`ߓ]xRγItSuGn5.M)hl2sDk`+Y|J!s(5;׼Y>isc_H#8O ʻ۹M0F\d1l 8F}Ṣ ˉ߿PNz.7Z/, w㔨CdRO9] [6k}"'?ǀާ'v6^~dU2m>x2;<SSЙ* ؅/ LjN&11@n[Y)>% |RM\qufD2/SOcHQ#Rb}ht QeWn:!@\H*R6ŧ`ND )OfƵ "Mh;y(>.RZ6|v|[04ҔQ_X#5條Vpe?4,V`d>tc׾0oVs##q3$|.DFqjrdwcNjӻn~}t%Yk'p?.G|09kIL9z5=e / M?]Urr=y P +AIUX{h !&Z.\ť7Oc) 28iR\ ['4D$0 rDMhb1ji%fBB$?^9y\)0T\WH#,^wD`"%f< W;mm,ÅXJsLLLȕ`X}(Nh1vFQ;va7$Pp S^~LZm3ps 6&#U'Z z 7fW<՝Eŏ6깘M|~9YOs)9 RU#es g-A9xAC@ Ov7Bn0S7)6 gG$ g2WfU01yr鬒Qj%7Y_V]Ŝd\K^ȸ9Oscx>ncX#0=du&7-[+zseا-\\+~8ׁ\!LݭN?0bEo̵bN埮sT(mUF1R0Gxyfn{r݀kY(wV]ݤQy#Vy/BO~ˈmmyIPMZ|IOeSs'oD"i|\nn f6 N-k^"/OV1єmt7U/B`ڸH2p[IN;[tWXM@ C4weAq=Yx!)u<[/) K P(G l^R5>LTo9譗$X}<[/)XXBpb%O[/ )Tl ǒSէLInv;z5W}EN;ey`qBUY@,(eQ_0!GE`t/.F8|`Ăf\Pk⑻7Ҹ`at\X2D}x[e^k\9o a9tS@Repε5BNku֬Cݞ{(-L%L恑1ɕ'(^=:B q,4ӈ:G VV,|M=4e1#t@!jM%Nˤ$>¼f/3tekYL$M_3:45SEF21Jqf%%:qfi@EAHI)R-]1~GҰQ4:gtȹp<+ɅJĐC.e, 8;\4hC R#riX[]RD ti$Ƅ&2 FR&!NK^0Is)!),ɱRBeMhrH_5c)Q 4&?ګ~i_Lr&X/΅e 4$_zzS/ Vځ}KOzS/ <фKCp;Q/ >'PzT/ `{ᗆ^ IzU/ Ƽ:v^/ ;)uj0ȞKbS3sHj sHKکv(9$)%f 6Vm]^В3ޗ*IBL~UL#tzT2 =фTL/)R_U*PwW٫TaxJ%R5!qR2)K4JT2 ^ex$Qԯ pᑀ?~ LH!iG \1 40i?&z=DHvH}HO`H4Q_ Ҝ,#}dx$+zᑀ`JT/>%H,>|(#JpIGX1 %檻$'aqE,\qLEw8OJ`бḶ|<Cad"%%8ճg%֜v!U>twJ y_J B*U)kD L`p?rqT$\@kMVc=l_,lb<50MY$>ll*mYr3cc#AS8l0?e0[ьa" zc\ ڱ/s)ҹ=\.FX`]0e"mkX AakHyx#jz vMz6?lJ 6$~Mr} PTx~9 xiMeRSY#ʯ5:ڮё( };W&~4|ڟfܷVMk[PHTƯW"k+ ֺF GYͳ72f!"+SoJ"@o5vb`|6n"sh#K# 1Yc'D b#ʄGF:w`< xz_1&)"ϳla}wf5wna\n3NSͦ~Vt5ewVU߾>W{f\bncaLXp?e]t3^Oޮ` y)|~]2~g/w2q{"Ow웷ǕcsWb8D&,7BQCKY$ߖ:ne/~<3WHw., '&q񼞘h"(m]ؖESiҕz_N W{֙d4d÷++]+.li1t2+2>|o/ j_As)h޲9^_gTPt A D=+^sd9N[Òp" 8 FE=rE(b rp* XCbwjQw뿍O# QRҩB\* 1xqD#J~B+ia5SV*(dek&gAqo,k6Oa*`dl&ۯ:j?70zr` OxMǓuQVuս#d\5d\{㚤!4^sn4Z/ij R<\)(Em;rvLF!lE7R')k ?w':6k}w5׍k .{|ss)mĤ4\(!Ҍ+ 7mH.dwlv=ݽq/;\z̭6׼{atzWA |[a~7<@Zm7G@]XS̓g׽xL4W_HT}y$P/|vmVN1'Z|~:_C[SNwX=ܹhc>)ډS~qtӏe{n]}Ch�[C^9ETR3jY\7}4l75Ä3WL ۽Q7ZgtR}O_OCз6owt}~P!ldӤ(0ypG=/Øb*k[s?ls[Me"ػKot)zo/ӗ8ݤd}vLLjOG2eްYw~]䏽OդzF쁅prX9@0(7a$JϦ|Y^u&'}4AnDn~$J8Q#j?<2Ra25-5((HJW<${"*U%':/XX˾yϳ+NAj. i'gO )ÂZz$-6pnwxH":qЊO/; H092NDzH7y tp^K'D&mOW};޲)Xj߮7U 5?$j~HN7W+ 罢,%((uP;)3HρL(toRk xO܌q'0 5`KXtB-ǟ'1GQ v7G7g_[lFUŬGjV\xPGliӣߌ^!*Ịu6<\nfLLqENӣNuN;ai=zM}odcC-]`S·B|ZB":%p6>$4trz㭖K6j{M:Da{@v8dyjݺ G;;ȻM72ELhn@n}0S H(Ctt}G6+bZ:uODˌn}0S Hg{0uAuMtpxѭ{{ft! Na&pN~󶞟RRJ25 EI1SI5֒%%žB5"bC()…w]j[}O{0Aͧ87rA8@rH(V&|PiXj9Sp_KPHNJM@-SR5Z")pp<9 ^$װWS/3G7F.nHJ VuK(2UKJQh~YX:e=[|H O?~_s:'=;}"RHK_il>q@\_sJ+)>҈QR\ƆVH J靠ԋ|{D(9|2\ے J%&~C a ȢC13Ni)e-d%)z}I* )ZE*!5C th0*EYI2i@go6ij"52BzcPyV+N :i$I@1SQ5ꍝ~7rA|oQNbkAUSrZ %fF_NiLP2 Lz]܈jCkv!$-˓⤷X^5*%)l IVs2,W-/zG4?O:~Xiir )#|Ҩ@ `+-=*P rAN[RڧHrdRZaac.ȑXlP&iեT5G(=|6Crn c$ZAALFscBIƸ@t\SP*øߛm&g6r'׋˴f3ӸA>L^AO4L"&P+f`^)&?YʜԐa;ݛf7teT~Ŀ|wco??q^dǟpF%/91 ?gQH ŝ+bEt*PDOCÕq@IǜYnaVvt5VVwvvM%1$|JuvƓtȉ|<{tJeHQhDI'w6?ⴢ{ $)IStI+h/lC6>\pZsC;Bmp:fuц{gXۇ>vlں'8FSW["̌V`o]\$_}nst̠(ʱ3t`\KGrYV: ZX:]:b(7mk7:ŁJGOq-{0]Α4]ՄcA:DFOXZ:!3e̚V\-ZBqZ*la"a@Y&Srt֌Igͷ)ͷOا2 |g>oXL_N8*d)WU<Ն+Y &`"ފKiFYu/z.K4V|OYsT\,X&jYXz,q:s>my8˅'Ҭs)*Ifҧ>oZe-eba)yKWRK籴Z3\biT< yB屴TTb)diR,jLizXz,Rކς\䱴.,=mc)RQ˝y[ e^z,# 1Zsf)eժ^`a3ss odsXcfNj]/FYuKb,}՚t"KXXKiz9?m_T#>$dȌՆKO9S5ו:d)R/_'"}w^J,f&?nBLԐjE4~C#zt;S÷`<+fqύO9%Ɯ]l^ pyРQ9XZzBdzBbW}saܺ3X7O غ oͣ W`FPf#h@3zl\9[F*pŊC4骋9DPnsrZP#LKS8ɜuRW*T3:Xm#;ojeTh5\Pp@8_4[}:MZyGEa Z(A6ǤBILHc[DRVc#a#s,@ZU#Bƅ Ma\i`1¥`[V Ԟs;v7}w?#3jjyҌ]\A+Bӷ.v#4us8'bqdk"5 c)m?xozӐ쉀h#^6+˫\G~]*WnSDq@IFPKi-|Tq T1yץ{DvH3 :oT,c@&Jme 57Ft5b$R=æ&Y5h-$Uઐ3yBɔpDΙ&2(mn`{P;?8[UęBpV2/c; ,wR/K'|jb7W\>)g_|=i:7>etTȅJf#fL36=)fY)T Z:V_2D%*ZyH(TA$=Z?{Wȍ /;6q%jӞac^zCQg7ْFNOHJbQ uZėD r;~zA 8tېvPG$y 0%R)Y gv UJ O}o$TJЇ?g\+^-|sr{9gh! B+~7st%wN?ܯV84}pң>|{+T$ǏPl7會4Hﯯ-۱Fɧi$4'fx~޽ G%P^ <8Z8qF<|m%9g$ G?j:8XiSh#UNf¹K8YTĮmT:=|OS>\=V5,UۧbT5{iC q٧ A1}yw.>hÉ :`4hp"Lԧ曊*wG%X+7RL~;nUFˆ- }w/A e-<ѻua!Dl2ՈػUMFnN;|ۘ*|b}{GօrٔcŴiI:9sJE(Ő͆zL~eeT.=uϣ 4?5Q[-rw_r~g<2[?0^{!"?wDbk2^6}JK./)Me6?x@72MJz bcgrT͡.QPf{ [׎G<f ¼4 ^g+VpEWP@ XPͤL$<|\ۣl7 k;1qw}07݊luuje7茋+O&% R ERӅpԅN&5p!/jänDt"W4(5DN^YJ"@(EgJIrIDƤ@_E $ZrQuS Tn".7?FG|\iYeEiMjZCm\TvQ܅T+K%2 QTҾ *"bQXF@#n:$twwĞhx<^9 &NtUrl) AZhPB20$SΒ!W)uy+g R yGI59pR@w@^礠xA]%ZjCr6缬tÛҌ"xJ ]hW B)}d9f0u+b;?XV$ ) S 9K4)bAؐB%w2<&.L"!1)A`0V"xĩ4IL qӋL9%4]>Zf^ͮyL8E|Qt\Y赦f^XA ]*KLʮ8 Mo~F ;Wʝ(pZ/yECs<)9׎5. Q6YcVkZ<A`8 ] u&>ގд̌|kX/tKKmX*ܩZԜU\È"5ÇQdQ:6(>92ta!DlQOu&9nL»bb:n#F=w ?vpn]X+7 l|L5˙^t`Vub/y jdbWputWosNH%}@bc/ҟj}\jTlۇ@Skp[:]alB] ^DeHP5*?7+kW4i^UMs@zYIrP$hsjJE @qc&62SD"8cwۛJ\/fm^buxX.WQﴆLͧ̂_^h2e7_8Z.),UpM$5 ɹ`H2{(m(Z4:VYjSN_A =ܞЇEN؝+ F #бxvQGPS!z4f{) N8D*TbH drI Ԓj1ͺG M$)/yQƔb8A GwegX4OH {V0#uRúr@3 6QJ` ̃蘕tA+=yv1J9-&2zLn9]xJ#'z{%QKjpAiֲ v&Ͷ/A{0Fa 3:ٵ A"H^sm$qW ƄC9Z FG է(B L9ʫ"i3ӓ-A6432QExCG ZN-e*nmAX@KI{2a;S# G؀ _%eۧwHn3خaגisYٳ、ɻtK:rb ,NszOP(OI/jچAr$EEKo(" vP#l/TL7V=U~J{W-YX8ZKRL~KYd|Lv@ Ů8WE]18reWB^ٔ0!tBƋNpܱhޭ y&bSGMH1 ˻0P>W:Wn5re&"Dm)N48q+ь<4ի2SGlyuGݗ8Z'%"_//>@S;#'%6Qp9c.on7?0^{GO/26fWMz;;|4<Hstu2*[}||wFuaJ||*w;jpVK%OD1TҾ *DLf4Lz+W'tBVĂI'F(_|Һ)#3Qf'B14J/XB\FSpU{]>OуRanW99;^yTN:M(U DK6||u2N b:$1@h R ^?ԄL6Wp5Er샊&tEr  dHyeMʚ5ͫiVI3 JŌ\!5 V .l^B&OO$͍3 `苂HC_ښ=T8z%$#yB T"MXɸ̲, tNXyR^D5/Wj8D99+Q;[A xю##:FCAaјB!b46y hI 2@a L""'CEB!"֬HSOF?Kh8bh" 2Fc$$+ .3&s٪GH,T3\$O)ee2еʱ*Q0 `=PDh!zUu[lQ^ ZK#$|rE C<7$D4)qCHhas0 a/8x҈bax:G w aiۮ&xJ)Qv=p"Z$G8Ǚ JS>3M;:a紒+;g{wf>=Op_ pZKSL+`:5yGly H5=e@Q7$jC]!yƫVI 6h/57`JzFZj#-)C K-t`y$}Q:T1Ղ[ ՂK T 9vYjjd{ިzZt-t d-*3o?c^Ճ6\%\78W3 +Q xݺ)k)7l'2 SEh^8W)ϔ2&(1 sʅȘ$Z"(wb~5}=uZ%dI}' LK4&A5]n\D[m߯7dm]\ %kty >˄+0/ۇf|6ן8}Q۝`d/ Z:k,p?~`|5}1Fta7JWT3nlz۸0W%3Q^z0AV,!$WvW[ilCl/'3C^ )&]h8=3GUS6%([U:AaƦ)[p6ڛPM1ęlWs8`í`Q[% P5IgJTRIR xFդ)͚RR+\RDQ591S7U|E$T13ϥBҬ)̌cGeT33~TM`LiƔdep]~QG!)oJ(b* J^U4oJQdԘ4bQ5ϥS_CW/IPF7)͏Rue"\zZ5Ҽ)DKcCՄn4oJQ(0hJ#QUv«)Ǔb`T13~TϔH)U}E3\U3S1Pm>wB ϥ\zڙ9fN)hK JA(T+QySRu^USr4GJy[tNLMJʖM m eSn]-\]wC-ÖR0VG88׻bδvh6 z.9i!cnA7My_ig÷<sgPλ|`P~jgX7xo#.GR\eNdwWynOͺ2ɘe0O( ׭'b}93ЛʥW4Lb7~@Z0bN>[wlT.M9i{gsFEpIPLJj20ΛRŜQ.$(Uģ4&1ϚR{+X4\wh$i u+Bk\V5 +ֵ[* tiuX+ [Hm~]TZMpg_ .1|var/home/core/zuul-output/logs/kubelet.log0000644000000000000000003703076715134375723017724 0ustar rootrootJan 22 09:41:55 crc systemd[1]: Starting Kubernetes Kubelet... Jan 22 09:41:55 crc restorecon[4758]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:55 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 22 09:41:56 crc restorecon[4758]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 22 09:41:56 crc kubenswrapper[4836]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.798155 4836 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802104 4836 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802124 4836 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802130 4836 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802136 4836 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802143 4836 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802150 4836 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802157 4836 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802164 4836 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802170 4836 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802175 4836 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802180 4836 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802185 4836 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802191 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802195 4836 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802200 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802207 4836 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802213 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802218 4836 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802224 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802229 4836 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802235 4836 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802241 4836 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802246 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802251 4836 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802256 4836 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802261 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802266 4836 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802271 4836 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802276 4836 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802281 4836 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802286 4836 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802291 4836 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802296 4836 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802302 4836 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802307 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802312 4836 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802316 4836 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802321 4836 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802327 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802332 4836 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802337 4836 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802342 4836 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802346 4836 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802351 4836 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802356 4836 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802360 4836 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802367 4836 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802372 4836 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802378 4836 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802382 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802387 4836 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802392 4836 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802396 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802401 4836 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802406 4836 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802411 4836 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802415 4836 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802420 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802425 4836 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802430 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802434 4836 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802439 4836 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802444 4836 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802448 4836 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802453 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802458 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802463 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802467 4836 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802472 4836 feature_gate.go:330] unrecognized feature gate: Example Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802479 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.802485 4836 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802748 4836 flags.go:64] FLAG: --address="0.0.0.0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802764 4836 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802774 4836 flags.go:64] FLAG: --anonymous-auth="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802782 4836 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802789 4836 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802794 4836 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802802 4836 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802809 4836 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802815 4836 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802821 4836 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802827 4836 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802833 4836 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802839 4836 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802844 4836 flags.go:64] FLAG: --cgroup-root="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802850 4836 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802856 4836 flags.go:64] FLAG: --client-ca-file="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802861 4836 flags.go:64] FLAG: --cloud-config="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802867 4836 flags.go:64] FLAG: --cloud-provider="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802872 4836 flags.go:64] FLAG: --cluster-dns="[]" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802879 4836 flags.go:64] FLAG: --cluster-domain="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802884 4836 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802890 4836 flags.go:64] FLAG: --config-dir="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802896 4836 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802902 4836 flags.go:64] FLAG: --container-log-max-files="5" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802909 4836 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802914 4836 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802920 4836 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802926 4836 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802932 4836 flags.go:64] FLAG: --contention-profiling="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802937 4836 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802943 4836 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802949 4836 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802954 4836 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802961 4836 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802968 4836 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802974 4836 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802979 4836 flags.go:64] FLAG: --enable-load-reader="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802985 4836 flags.go:64] FLAG: --enable-server="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802991 4836 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.802998 4836 flags.go:64] FLAG: --event-burst="100" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803004 4836 flags.go:64] FLAG: --event-qps="50" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803010 4836 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803015 4836 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803021 4836 flags.go:64] FLAG: --eviction-hard="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803028 4836 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803050 4836 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803057 4836 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803062 4836 flags.go:64] FLAG: --eviction-soft="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803068 4836 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803074 4836 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803080 4836 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803085 4836 flags.go:64] FLAG: --experimental-mounter-path="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803091 4836 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803096 4836 flags.go:64] FLAG: --fail-swap-on="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803102 4836 flags.go:64] FLAG: --feature-gates="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803109 4836 flags.go:64] FLAG: --file-check-frequency="20s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803114 4836 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803120 4836 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803127 4836 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803133 4836 flags.go:64] FLAG: --healthz-port="10248" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803138 4836 flags.go:64] FLAG: --help="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803144 4836 flags.go:64] FLAG: --hostname-override="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803150 4836 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803155 4836 flags.go:64] FLAG: --http-check-frequency="20s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803161 4836 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803167 4836 flags.go:64] FLAG: --image-credential-provider-config="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803172 4836 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803178 4836 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803183 4836 flags.go:64] FLAG: --image-service-endpoint="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803189 4836 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803195 4836 flags.go:64] FLAG: --kube-api-burst="100" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803201 4836 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803207 4836 flags.go:64] FLAG: --kube-api-qps="50" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803212 4836 flags.go:64] FLAG: --kube-reserved="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803218 4836 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803223 4836 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803229 4836 flags.go:64] FLAG: --kubelet-cgroups="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803234 4836 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803240 4836 flags.go:64] FLAG: --lock-file="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803245 4836 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803251 4836 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803256 4836 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803264 4836 flags.go:64] FLAG: --log-json-split-stream="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803270 4836 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803275 4836 flags.go:64] FLAG: --log-text-split-stream="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803281 4836 flags.go:64] FLAG: --logging-format="text" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803286 4836 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803293 4836 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803298 4836 flags.go:64] FLAG: --manifest-url="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803303 4836 flags.go:64] FLAG: --manifest-url-header="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803315 4836 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803321 4836 flags.go:64] FLAG: --max-open-files="1000000" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803327 4836 flags.go:64] FLAG: --max-pods="110" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803333 4836 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803339 4836 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803345 4836 flags.go:64] FLAG: --memory-manager-policy="None" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803350 4836 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803357 4836 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803362 4836 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803368 4836 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803380 4836 flags.go:64] FLAG: --node-status-max-images="50" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803386 4836 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803391 4836 flags.go:64] FLAG: --oom-score-adj="-999" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803397 4836 flags.go:64] FLAG: --pod-cidr="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803402 4836 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803411 4836 flags.go:64] FLAG: --pod-manifest-path="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803417 4836 flags.go:64] FLAG: --pod-max-pids="-1" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803423 4836 flags.go:64] FLAG: --pods-per-core="0" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803428 4836 flags.go:64] FLAG: --port="10250" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803434 4836 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803440 4836 flags.go:64] FLAG: --provider-id="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803446 4836 flags.go:64] FLAG: --qos-reserved="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803452 4836 flags.go:64] FLAG: --read-only-port="10255" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803457 4836 flags.go:64] FLAG: --register-node="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803463 4836 flags.go:64] FLAG: --register-schedulable="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803468 4836 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803477 4836 flags.go:64] FLAG: --registry-burst="10" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803483 4836 flags.go:64] FLAG: --registry-qps="5" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803488 4836 flags.go:64] FLAG: --reserved-cpus="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803494 4836 flags.go:64] FLAG: --reserved-memory="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803500 4836 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803506 4836 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803512 4836 flags.go:64] FLAG: --rotate-certificates="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803518 4836 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803523 4836 flags.go:64] FLAG: --runonce="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803529 4836 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803534 4836 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803540 4836 flags.go:64] FLAG: --seccomp-default="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803546 4836 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803551 4836 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803558 4836 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803563 4836 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803569 4836 flags.go:64] FLAG: --storage-driver-password="root" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803576 4836 flags.go:64] FLAG: --storage-driver-secure="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803582 4836 flags.go:64] FLAG: --storage-driver-table="stats" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803588 4836 flags.go:64] FLAG: --storage-driver-user="root" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803594 4836 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803600 4836 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803606 4836 flags.go:64] FLAG: --system-cgroups="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803612 4836 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803621 4836 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803626 4836 flags.go:64] FLAG: --tls-cert-file="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803632 4836 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803639 4836 flags.go:64] FLAG: --tls-min-version="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803645 4836 flags.go:64] FLAG: --tls-private-key-file="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803650 4836 flags.go:64] FLAG: --topology-manager-policy="none" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803656 4836 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803662 4836 flags.go:64] FLAG: --topology-manager-scope="container" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803667 4836 flags.go:64] FLAG: --v="2" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803674 4836 flags.go:64] FLAG: --version="false" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803681 4836 flags.go:64] FLAG: --vmodule="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803688 4836 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.803694 4836 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803823 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803830 4836 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803836 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803841 4836 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803846 4836 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803851 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803856 4836 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803862 4836 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803866 4836 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803872 4836 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803876 4836 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803883 4836 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803889 4836 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803895 4836 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803900 4836 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803906 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803911 4836 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803917 4836 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803923 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803928 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803933 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803938 4836 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803943 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803947 4836 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803953 4836 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803958 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803963 4836 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803968 4836 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803973 4836 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803978 4836 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803982 4836 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803987 4836 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803992 4836 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.803997 4836 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804002 4836 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804007 4836 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804012 4836 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804017 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804025 4836 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804030 4836 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804035 4836 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804055 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804060 4836 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804065 4836 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804070 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804075 4836 feature_gate.go:330] unrecognized feature gate: Example Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804080 4836 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804084 4836 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804090 4836 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804094 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804100 4836 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804105 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804109 4836 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804114 4836 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804119 4836 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804124 4836 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804128 4836 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804133 4836 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804138 4836 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804143 4836 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804147 4836 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804153 4836 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804158 4836 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804165 4836 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804171 4836 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804176 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804181 4836 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804186 4836 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804191 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804198 4836 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.804206 4836 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.804221 4836 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.813409 4836 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.813451 4836 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813600 4836 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813621 4836 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813634 4836 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813647 4836 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813655 4836 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813664 4836 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813673 4836 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813681 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813689 4836 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813697 4836 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813705 4836 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813713 4836 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813720 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813728 4836 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813736 4836 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813744 4836 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813751 4836 feature_gate.go:330] unrecognized feature gate: Example Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813759 4836 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813767 4836 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813775 4836 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813782 4836 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813790 4836 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813797 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813805 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813814 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813822 4836 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813832 4836 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813843 4836 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813854 4836 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813865 4836 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813874 4836 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813883 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813890 4836 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813902 4836 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813913 4836 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813922 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813931 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813940 4836 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813951 4836 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813961 4836 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813970 4836 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813978 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813986 4836 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.813993 4836 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814001 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814010 4836 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814018 4836 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814026 4836 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814072 4836 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814082 4836 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814091 4836 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814098 4836 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814106 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814114 4836 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814122 4836 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814133 4836 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814143 4836 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814152 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814161 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814169 4836 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814177 4836 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814186 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814193 4836 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814205 4836 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814213 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814221 4836 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814229 4836 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814237 4836 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814244 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814252 4836 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814260 4836 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.814274 4836 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814526 4836 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814541 4836 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814552 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814561 4836 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814570 4836 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814577 4836 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814585 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814594 4836 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814602 4836 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814613 4836 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814624 4836 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814633 4836 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814641 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814650 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814659 4836 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814668 4836 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814676 4836 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814684 4836 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814692 4836 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814701 4836 feature_gate.go:330] unrecognized feature gate: Example Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814709 4836 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814717 4836 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814724 4836 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814732 4836 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814740 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814750 4836 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814760 4836 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814768 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814777 4836 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814786 4836 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814795 4836 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814803 4836 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814811 4836 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814819 4836 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814827 4836 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814835 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814844 4836 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814853 4836 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814862 4836 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814869 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814880 4836 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814890 4836 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814898 4836 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814906 4836 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814916 4836 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814928 4836 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814940 4836 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814949 4836 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814959 4836 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814968 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814978 4836 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.814990 4836 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815000 4836 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815010 4836 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815017 4836 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815025 4836 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815034 4836 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815072 4836 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815080 4836 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815088 4836 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815096 4836 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815104 4836 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815111 4836 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815119 4836 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815126 4836 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815135 4836 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815143 4836 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815150 4836 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815159 4836 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815169 4836 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.815178 4836 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.815193 4836 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.815820 4836 server.go:940] "Client rotation is on, will bootstrap in background" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.821686 4836 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.821858 4836 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.823005 4836 server.go:997] "Starting client certificate rotation" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.823087 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.823321 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 10:47:56.826062447 +0000 UTC Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.823422 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.831698 4836 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.833507 4836 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.837583 4836 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.848902 4836 log.go:25] "Validated CRI v1 runtime API" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.875884 4836 log.go:25] "Validated CRI v1 image API" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.878887 4836 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.881779 4836 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-22-09-37-00-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.881830 4836 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.908003 4836 manager.go:217] Machine: {Timestamp:2026-01-22 09:41:56.905935392 +0000 UTC m=+0.265703989 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ed462382-29a2-4539-9f5b-0dbc69ba727c BootID:cfbd5c21-5034-481d-a75e-e8d2d8dd038d Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ab:ff:be Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ab:ff:be Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e1:87:19 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a8:71:b4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f2:37:3f Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2c:df:2f Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:31:a0:3e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ba:4b:4d:7a:35:9e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:af:78:eb:25:84 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.908458 4836 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.908779 4836 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.912316 4836 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.913226 4836 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.913289 4836 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.913688 4836 topology_manager.go:138] "Creating topology manager with none policy" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.913712 4836 container_manager_linux.go:303] "Creating device plugin manager" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.913982 4836 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.914065 4836 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.914539 4836 state_mem.go:36] "Initialized new in-memory state store" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.914697 4836 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.915818 4836 kubelet.go:418] "Attempting to sync node with API server" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.915857 4836 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.915926 4836 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.915950 4836 kubelet.go:324] "Adding apiserver pod source" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.915970 4836 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.918404 4836 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.918919 4836 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.920624 4836 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.921284 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.921292 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.921394 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.921409 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921537 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921574 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921598 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921612 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921645 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921660 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921683 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921717 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921732 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921749 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921776 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.921799 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.922152 4836 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.922900 4836 server.go:1280] "Started kubelet" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.923202 4836 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.923579 4836 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.923615 4836 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 22 09:41:56 crc systemd[1]: Started Kubernetes Kubelet. Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.924643 4836 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.926994 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.927120 4836 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.927566 4836 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.927615 4836 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.927524 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 13:11:45.841084401 +0000 UTC Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.927636 4836 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.927686 4836 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.928785 4836 factory.go:55] Registering systemd factory Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.928838 4836 factory.go:221] Registration of the systemd container factory successfully Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.929021 4836 server.go:460] "Adding debug handlers to kubelet server" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.929959 4836 factory.go:153] Registering CRI-O factory Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.930213 4836 factory.go:221] Registration of the crio container factory successfully Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.930123 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188d0443943d8b42 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:41:56.92285421 +0000 UTC m=+0.282622797,LastTimestamp:2026-01-22 09:41:56.92285421 +0000 UTC m=+0.282622797,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.930989 4836 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.931161 4836 factory.go:103] Registering Raw factory Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.931182 4836 manager.go:1196] Started watching for new ooms in manager Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.931644 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="200ms" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.932906 4836 manager.go:319] Starting recovery of all containers Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.938892 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.939032 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.943935 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188d0443943d8b42 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:41:56.92285421 +0000 UTC m=+0.282622797,LastTimestamp:2026-01-22 09:41:56.92285421 +0000 UTC m=+0.282622797,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.953923 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954033 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954093 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954121 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954157 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954190 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954266 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954287 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954313 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954334 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954356 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954378 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954402 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954427 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954446 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954464 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954490 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954508 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954527 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954547 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954566 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954583 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954603 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954621 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954640 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954660 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954682 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954706 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954725 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954751 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954769 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954853 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954871 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954891 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954907 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954926 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954944 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.954962 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955014 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955035 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955092 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955117 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955139 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955171 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955192 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955210 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955229 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955247 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955266 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955284 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955306 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955328 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955354 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955375 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955394 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955417 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955436 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955455 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955473 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955490 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955507 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955526 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955543 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955563 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955581 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955600 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955620 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955648 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955669 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955685 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955702 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955722 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955739 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955757 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955780 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955799 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955816 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955835 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955853 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955871 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955888 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955907 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955929 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955958 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955977 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.955994 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956013 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956030 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956081 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956105 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956124 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956142 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956162 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956179 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956196 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956213 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956232 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956251 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956271 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956329 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956348 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956366 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956384 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956403 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956430 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956450 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956469 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956490 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956512 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956534 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956553 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956573 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956607 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956628 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956650 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956670 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956687 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956707 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956727 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956744 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956762 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956779 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956797 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956816 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956834 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956853 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956869 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956888 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956906 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956926 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956946 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956963 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.956993 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957010 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957029 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957088 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957116 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957144 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957202 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957222 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957241 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957263 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957283 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957304 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957323 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957344 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957374 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957392 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957411 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957440 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957459 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957476 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957495 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957513 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957532 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957550 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957568 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957587 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957604 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957623 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957643 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957661 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957700 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957717 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.957735 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958671 4836 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958713 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958735 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958754 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958773 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958834 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958859 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958877 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958894 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958912 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958931 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958949 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.958966 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959018 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959036 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959097 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959124 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959147 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959165 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959184 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959201 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959252 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959271 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959356 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959374 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959394 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959412 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959431 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959449 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959468 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959485 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959507 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959526 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959543 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959562 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959579 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959596 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959614 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959636 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959655 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959672 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959689 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959706 4836 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959725 4836 reconstruct.go:97] "Volume reconstruction finished" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.959738 4836 reconciler.go:26] "Reconciler: start to sync state" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.966594 4836 manager.go:324] Recovery completed Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.974284 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.976415 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.976452 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.976462 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.977824 4836 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.977864 4836 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.977896 4836 state_mem.go:36] "Initialized new in-memory state store" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.986333 4836 policy_none.go:49] "None policy: Start" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.987575 4836 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.987643 4836 state_mem.go:35] "Initializing new in-memory state store" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.991853 4836 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.995746 4836 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.995816 4836 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 22 09:41:56 crc kubenswrapper[4836]: I0122 09:41:56.995859 4836 kubelet.go:2335] "Starting kubelet main sync loop" Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.995936 4836 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 22 09:41:56 crc kubenswrapper[4836]: W0122 09:41:56.997196 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:56 crc kubenswrapper[4836]: E0122 09:41:56.997304 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.027722 4836 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.060664 4836 manager.go:334] "Starting Device Plugin manager" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.060783 4836 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.060806 4836 server.go:79] "Starting device plugin registration server" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.061479 4836 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.061709 4836 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.062145 4836 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.062335 4836 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.062751 4836 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.072676 4836 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.096102 4836 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.096259 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.097773 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.097823 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.097838 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.097975 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.098605 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.098691 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.099574 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.099623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.099644 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.099819 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.100135 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.100190 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.100258 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.100291 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.100307 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.101289 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.101328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.101345 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.102386 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.102755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.102796 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.102847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.103243 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.103298 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.104393 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.105367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.105395 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106198 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106265 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106296 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106327 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106380 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.106327 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108602 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108660 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108815 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.108923 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.109865 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.109912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.109932 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.132381 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="400ms" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162129 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162281 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162344 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162384 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162418 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162455 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162490 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162566 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162731 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162806 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162908 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162937 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162956 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162979 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.162994 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.163007 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.164188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.164244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.164265 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.164301 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.164868 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.173:6443: connect: connection refused" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264314 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264422 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264519 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264593 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264625 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264702 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264642 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264666 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264690 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264720 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264795 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264953 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264988 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264991 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265072 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265006 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265139 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265180 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265191 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.264639 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265244 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265320 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265417 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265432 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265505 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265542 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265594 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265605 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265730 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.265635 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.365672 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.367832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.367880 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.367898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.367930 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.368396 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.173:6443: connect: connection refused" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.432711 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.442004 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.463631 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: W0122 09:41:57.470288 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-11018c9c67a85b0a1e661bafb5a97edab592dfa85d8a82955552f769e57fb58f WatchSource:0}: Error finding container 11018c9c67a85b0a1e661bafb5a97edab592dfa85d8a82955552f769e57fb58f: Status 404 returned error can't find the container with id 11018c9c67a85b0a1e661bafb5a97edab592dfa85d8a82955552f769e57fb58f Jan 22 09:41:57 crc kubenswrapper[4836]: W0122 09:41:57.472600 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ac8041968e58825f318be8df2c2066143714d725762e015f43138f289a1593da WatchSource:0}: Error finding container ac8041968e58825f318be8df2c2066143714d725762e015f43138f289a1593da: Status 404 returned error can't find the container with id ac8041968e58825f318be8df2c2066143714d725762e015f43138f289a1593da Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.472883 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.478080 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:41:57 crc kubenswrapper[4836]: W0122 09:41:57.488432 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-af50ca53425f56a413ac462ee4423aa8af950b3605c16a845a44474d61d6e5a7 WatchSource:0}: Error finding container af50ca53425f56a413ac462ee4423aa8af950b3605c16a845a44474d61d6e5a7: Status 404 returned error can't find the container with id af50ca53425f56a413ac462ee4423aa8af950b3605c16a845a44474d61d6e5a7 Jan 22 09:41:57 crc kubenswrapper[4836]: W0122 09:41:57.497432 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-40f011838f698f924f2496f692c5628eb34a97529e492f89a7bdefdb90271eae WatchSource:0}: Error finding container 40f011838f698f924f2496f692c5628eb34a97529e492f89a7bdefdb90271eae: Status 404 returned error can't find the container with id 40f011838f698f924f2496f692c5628eb34a97529e492f89a7bdefdb90271eae Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.533238 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="800ms" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.768782 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.770650 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.770703 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.770715 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.770743 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: E0122 09:41:57.771242 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.173:6443: connect: connection refused" node="crc" Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.924067 4836 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:57 crc kubenswrapper[4836]: I0122 09:41:57.928124 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 22:45:45.622728642 +0000 UTC Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.003086 4836 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269" exitCode=0 Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.003248 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.003426 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"11018c9c67a85b0a1e661bafb5a97edab592dfa85d8a82955552f769e57fb58f"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.003635 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.005147 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.005201 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.005222 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.005300 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.005351 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ac8041968e58825f318be8df2c2066143714d725762e015f43138f289a1593da"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.006660 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c" exitCode=0 Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.006725 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.006764 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5692d7216f427da094f3946a15f0843084e10b426c4d56fea038d0248b369704"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.006889 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.007948 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.007983 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.007996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.009138 4836 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b09c906f259b78a035978188ec2703dbbbf30c702c39192bb93762c94c4e710d" exitCode=0 Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.009218 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b09c906f259b78a035978188ec2703dbbbf30c702c39192bb93762c94c4e710d"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.009245 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"40f011838f698f924f2496f692c5628eb34a97529e492f89a7bdefdb90271eae"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.009339 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.009350 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010402 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010430 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010447 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010443 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.010588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.011639 4836 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020" exitCode=0 Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.011669 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.011707 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"af50ca53425f56a413ac462ee4423aa8af950b3605c16a845a44474d61d6e5a7"} Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.011786 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.012542 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.012572 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.012584 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: W0122 09:41:58.107383 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.107481 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:58 crc kubenswrapper[4836]: W0122 09:41:58.123415 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.123484 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:58 crc kubenswrapper[4836]: W0122 09:41:58.204599 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.204697 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.334407 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="1.6s" Jan 22 09:41:58 crc kubenswrapper[4836]: W0122 09:41:58.505371 4836 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.173:6443: connect: connection refused Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.505765 4836 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.173:6443: connect: connection refused" logger="UnhandledError" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.572130 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.573548 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.573595 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.573607 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.573637 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:41:58 crc kubenswrapper[4836]: E0122 09:41:58.574190 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.173:6443: connect: connection refused" node="crc" Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.859689 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 09:41:58 crc kubenswrapper[4836]: I0122 09:41:58.929080 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 01:43:50.963583623 +0000 UTC Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.017387 4836 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="49e1836d9dca0131a7fe5d6dbad3c4ac82547a3e24d5dbc8e1ae67e1ca58cb5f" exitCode=0 Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.017487 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"49e1836d9dca0131a7fe5d6dbad3c4ac82547a3e24d5dbc8e1ae67e1ca58cb5f"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.017747 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.019562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.019604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.019614 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.021766 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.021995 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.024096 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.024208 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.024257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.038636 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.039176 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.039222 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.039236 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.041475 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.041510 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.041525 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.047295 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.047330 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.047344 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.047515 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.048465 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.048495 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.048507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.051587 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.051616 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.051629 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.051640 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6"} Jan 22 09:41:59 crc kubenswrapper[4836]: I0122 09:41:59.929805 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:45:41.869597711 +0000 UTC Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.058086 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7"} Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.058294 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.059666 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.059733 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.059760 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063128 4836 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2882b8881648a15f8e59ec80b8d7eb6609ddef26273a241fbdd42cb2c9516031" exitCode=0 Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063181 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2882b8881648a15f8e59ec80b8d7eb6609ddef26273a241fbdd42cb2c9516031"} Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063280 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063321 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063412 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.063331 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065169 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065218 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065234 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065347 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.065375 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.066364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.066418 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.066441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.174916 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.176513 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.176569 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.176586 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.176621 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:42:00 crc kubenswrapper[4836]: I0122 09:42:00.930899 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 09:35:56.512831709 +0000 UTC Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.075419 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cbc3693297d00876fb076d1932465ecc4d68fef092c1f22c575be60756f25ee2"} Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.075514 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c18fb67587ee25426c9e329597484447ed487d65744fdc0c49a6c783fcd2675"} Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.075546 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bba5861aff1c8c5a7ebc91786bc68318feab9270b6adddc13370e523ba7f9408"} Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.075459 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.075642 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.080574 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.080658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.080931 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.742133 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.742410 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.744114 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.744189 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.744215 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.931859 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 06:51:35.523027995 +0000 UTC Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.937523 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.944566 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:01 crc kubenswrapper[4836]: I0122 09:42:01.986642 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.084105 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"27ffd29b6cff6ee2de182539e84163e4201a5c520afbb48a378cee7830b327fd"} Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.084187 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.084208 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0d1ecf3e5ce5d6f3f48573a9f0c86d623d73becc0e3a8942d7338815934ac935"} Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.084155 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.085747 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.085803 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.085822 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.086024 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.086135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.086155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.583465 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.583712 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.583782 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.585254 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.585339 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.585367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:02 crc kubenswrapper[4836]: I0122 09:42:02.933004 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 19:24:20.523759698 +0000 UTC Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.019663 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.086605 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.086644 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088231 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088346 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.088299 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.642314 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.642603 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.644197 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.644275 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.644297 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.806469 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:03 crc kubenswrapper[4836]: I0122 09:42:03.933508 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 03:30:22.699100276 +0000 UTC Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.064346 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.089333 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.089333 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.089340 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.090885 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.090962 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.090993 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091210 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091267 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091532 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091606 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.091628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.390651 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.743214 4836 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.743624 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 22 09:42:04 crc kubenswrapper[4836]: I0122 09:42:04.934408 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 10:54:57.131371924 +0000 UTC Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.091894 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.092924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.092988 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.093007 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.448772 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.449033 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.450820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.450899 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.450922 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:05 crc kubenswrapper[4836]: I0122 09:42:05.935216 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 03:09:53.834735599 +0000 UTC Jan 22 09:42:06 crc kubenswrapper[4836]: I0122 09:42:06.935961 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 05:21:36.878793849 +0000 UTC Jan 22 09:42:07 crc kubenswrapper[4836]: E0122 09:42:07.072822 4836 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 22 09:42:07 crc kubenswrapper[4836]: I0122 09:42:07.937000 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 07:52:55.818886993 +0000 UTC Jan 22 09:42:08 crc kubenswrapper[4836]: E0122 09:42:08.862470 4836 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 22 09:42:08 crc kubenswrapper[4836]: I0122 09:42:08.925224 4836 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 22 09:42:08 crc kubenswrapper[4836]: I0122 09:42:08.937618 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 00:30:11.157099539 +0000 UTC Jan 22 09:42:09 crc kubenswrapper[4836]: E0122 09:42:09.935904 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Jan 22 09:42:09 crc kubenswrapper[4836]: I0122 09:42:09.938121 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 19:13:31.763422863 +0000 UTC Jan 22 09:42:10 crc kubenswrapper[4836]: E0122 09:42:10.178426 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Jan 22 09:42:10 crc kubenswrapper[4836]: I0122 09:42:10.227239 4836 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 22 09:42:10 crc kubenswrapper[4836]: I0122 09:42:10.227330 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 22 09:42:10 crc kubenswrapper[4836]: I0122 09:42:10.237459 4836 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 22 09:42:10 crc kubenswrapper[4836]: I0122 09:42:10.237957 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 22 09:42:10 crc kubenswrapper[4836]: I0122 09:42:10.938708 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 12:52:10.602770535 +0000 UTC Jan 22 09:42:11 crc kubenswrapper[4836]: I0122 09:42:11.939740 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 12:19:20.138337166 +0000 UTC Jan 22 09:42:12 crc kubenswrapper[4836]: I0122 09:42:12.940432 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 12:50:58.573364454 +0000 UTC Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.027730 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.028261 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.030158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.030241 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.030264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.231895 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.249640 4836 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.379139 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.380796 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.380858 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.380876 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.380911 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:42:13 crc kubenswrapper[4836]: E0122 09:42:13.387281 4836 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.817181 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.817440 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.819986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.820109 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.820139 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.824955 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:13 crc kubenswrapper[4836]: I0122 09:42:13.940847 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 10:08:25.605463682 +0000 UTC Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.099518 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.099694 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.100852 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.100900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.100914 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.116800 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.116957 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.117641 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.117687 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.117955 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.118237 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.118340 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.122754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.123005 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.123181 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.157549 4836 csr.go:261] certificate signing request csr-42p6t is approved, waiting to be issued Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.178342 4836 csr.go:257] certificate signing request csr-42p6t is issued Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.743174 4836 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.743276 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 09:42:14 crc kubenswrapper[4836]: I0122 09:42:14.940966 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 11:28:46.140099242 +0000 UTC Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.179612 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-22 09:37:14 +0000 UTC, rotation deadline is 2026-11-15 20:26:27.660193921 +0000 UTC Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.179659 4836 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7138h44m12.480538282s for next certificate rotation Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.225922 4836 trace.go:236] Trace[1310537542]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:01.080) (total time: 14145ms): Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1310537542]: ---"Objects listed" error: 14145ms (09:42:15.225) Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1310537542]: [14.14514238s] [14.14514238s] END Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.225964 4836 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.225999 4836 trace.go:236] Trace[301132006]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:01.379) (total time: 13846ms): Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[301132006]: ---"Objects listed" error: 13845ms (09:42:15.225) Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[301132006]: [13.846032621s] [13.846032621s] END Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.226031 4836 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.226815 4836 trace.go:236] Trace[1356654680]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:00.795) (total time: 14431ms): Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1356654680]: ---"Objects listed" error: 14431ms (09:42:15.226) Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1356654680]: [14.431495078s] [14.431495078s] END Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.226850 4836 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.228432 4836 trace.go:236] Trace[1449549292]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:00.535) (total time: 14693ms): Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1449549292]: ---"Objects listed" error: 14692ms (09:42:15.228) Jan 22 09:42:15 crc kubenswrapper[4836]: Trace[1449549292]: [14.693008632s] [14.693008632s] END Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.228746 4836 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.229097 4836 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.593171 4836 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34488->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.593486 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:34488->192.168.126.11:17697: read: connection reset by peer" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.594034 4836 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.594118 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.928636 4836 apiserver.go:52] "Watching apiserver" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.931623 4836 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.931943 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.932388 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.932406 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.932385 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:15 crc kubenswrapper[4836]: E0122 09:42:15.932490 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.932411 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:15 crc kubenswrapper[4836]: E0122 09:42:15.932561 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.932588 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.933828 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:15 crc kubenswrapper[4836]: E0122 09:42:15.933906 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.934303 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.938974 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939000 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939187 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939300 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939340 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939526 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939552 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.939561 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 09:42:15 crc kubenswrapper[4836]: I0122 09:42:15.941721 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 04:21:59.057588336 +0000 UTC Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.029349 4836 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033769 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033806 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033831 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033856 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033876 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033898 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033913 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033931 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033951 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033969 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.033987 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034005 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034021 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034051 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034068 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034086 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034103 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034123 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034143 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034160 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034177 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034193 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034209 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034225 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034215 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034242 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034259 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034297 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034315 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034331 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034350 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034366 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034381 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034398 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034417 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034433 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034453 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034472 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034528 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034545 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034561 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034577 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034544 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034639 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034653 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034594 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034726 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034750 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034770 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034788 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034805 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034821 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034839 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034856 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034864 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034874 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034874 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034871 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034924 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034946 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034965 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034981 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.034998 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035016 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035034 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035066 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035069 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035083 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035100 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035116 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035158 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035181 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035197 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035211 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035229 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035245 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035261 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035276 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035292 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035309 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035326 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035341 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035355 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035375 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035391 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035407 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035458 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035475 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035502 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035518 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035534 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035548 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035564 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035583 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035599 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035615 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035630 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035644 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035661 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035678 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035694 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035712 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035730 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035744 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035793 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035813 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035831 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035851 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035871 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035888 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035907 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035947 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035964 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035982 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035998 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036015 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036032 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036069 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036087 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036104 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036119 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036135 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036154 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036170 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036187 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036204 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036222 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036305 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036325 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036341 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036357 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036374 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036388 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036404 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036422 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036439 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036457 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036475 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036501 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036518 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036537 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036554 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036570 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036589 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036608 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036624 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036642 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036659 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036677 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036696 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036712 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036736 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036755 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036771 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036788 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036805 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036822 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036837 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036853 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036894 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036919 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036938 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036955 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036972 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036988 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037005 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037025 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037057 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037074 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037091 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037107 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037124 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037141 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037156 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037174 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037192 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037208 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037225 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037241 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037257 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037275 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037291 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037309 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037327 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037344 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037362 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037379 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037396 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037414 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037431 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037446 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037463 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037482 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037499 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037515 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037531 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037558 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037575 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037595 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037613 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037631 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037648 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037665 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037709 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037732 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037754 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037773 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037802 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037821 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037838 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037858 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037878 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037900 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037920 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037940 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037959 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037979 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038031 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038357 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038369 4836 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038380 4836 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038392 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038402 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038413 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035307 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035314 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035424 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035436 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035486 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035607 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035634 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035648 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040478 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035758 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035817 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035830 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035842 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035954 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.035996 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036001 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036136 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036153 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036188 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036220 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036303 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036436 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036453 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036529 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036616 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036624 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036677 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036761 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036824 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036854 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036892 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.036988 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037293 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037305 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037390 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037450 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037534 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037581 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037595 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037685 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037701 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037746 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037782 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037841 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037901 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037910 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.037943 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038129 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038300 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038477 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038604 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038918 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038930 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.038985 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040260 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040367 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040696 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040752 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.040105 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.041137 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.041586 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.041734 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.041956 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.041960 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.042180 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.045839 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.046927 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.047077 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.047434 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.049373 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.050140 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.050304 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.052191 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.052488 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.052779 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.053140 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.053505 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.053785 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.054062 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.054114 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.054612 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.061287 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.062884 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063126 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063305 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063434 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063463 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063613 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063640 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063663 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.063994 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.064276 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.064922 4836 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.065268 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.066283 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.066403 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.066744 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.066960 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067140 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067314 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067565 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067616 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067800 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067841 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.067989 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068355 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068569 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068609 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068765 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068817 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.068871 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069029 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069162 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069173 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069402 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069837 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.070222 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.069067 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.075472 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.075769 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.075966 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.075974 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.076177 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.076392 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.076533 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.082306 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.082454 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.082528 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:16.582501912 +0000 UTC m=+19.942270429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.082828 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.082916 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083106 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083217 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083300 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083497 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083679 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083689 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.083959 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.084580 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.085337 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.085347 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.086351 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.086436 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.086494 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:16.586472065 +0000 UTC m=+19.946240772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.087145 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.087476 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.088113 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.088746 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.089377 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.090443 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.091171 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.091706 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.092108 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.099413 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.102310 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.102504 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.102685 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.102758 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.102920 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.103550 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.104312 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.105417 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.105654 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.105685 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.105716 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:16.605697127 +0000 UTC m=+19.965465824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.106014 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.106031 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.106278 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.106412 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.106740 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.109157 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.109699 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110438 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110598 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110785 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110839 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110890 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.110866 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.111279 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.111376 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.112029 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.113354 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.113594 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.114191 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.115549 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.118172 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.118419 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.118901 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.119012 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.119999 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.120407 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.120431 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.120575 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.120776 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.121597 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.123934 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.124628 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.132898 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133396 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133431 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133445 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133511 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:16.633491323 +0000 UTC m=+19.993259840 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133574 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133585 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133592 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.133614 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:16.633607956 +0000 UTC m=+19.993376473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.133949 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138736 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138792 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138840 4836 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138850 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138859 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138868 4836 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138877 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138885 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138894 4836 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138901 4836 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138911 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138922 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138932 4836 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138943 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138955 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138967 4836 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138978 4836 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.138989 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139000 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139011 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139022 4836 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139058 4836 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139074 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139086 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139099 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139111 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139124 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139147 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139156 4836 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139164 4836 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139174 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139183 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139194 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139203 4836 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139212 4836 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139220 4836 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139230 4836 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139238 4836 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139247 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139257 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139266 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139274 4836 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139283 4836 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139291 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139300 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139310 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139321 4836 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139330 4836 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139339 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139348 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139356 4836 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139367 4836 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139376 4836 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139389 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139398 4836 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139406 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139416 4836 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139425 4836 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139438 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139448 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139459 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139469 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139477 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139485 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139493 4836 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139501 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139510 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139517 4836 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139526 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139534 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139542 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.139550 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.141875 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.141963 4836 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.141987 4836 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.141998 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142009 4836 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142018 4836 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142027 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142051 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142060 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142079 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142087 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142096 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142104 4836 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142116 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142126 4836 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142137 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142151 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142160 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142168 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142177 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142186 4836 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142196 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.141996 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142205 4836 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142230 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142255 4836 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142265 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142274 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142283 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142293 4836 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142302 4836 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142310 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142320 4836 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142328 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142337 4836 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142345 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142354 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142363 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142371 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142380 4836 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142389 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142398 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142408 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142420 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142434 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142449 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142463 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142475 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142487 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142502 4836 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142512 4836 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142521 4836 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142530 4836 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142539 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142548 4836 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142557 4836 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142567 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142576 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142586 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142596 4836 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142605 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142617 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142626 4836 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142635 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142645 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142655 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142665 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142673 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142682 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142690 4836 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142699 4836 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142712 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142721 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142729 4836 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142738 4836 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142747 4836 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142755 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142763 4836 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142771 4836 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142780 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142790 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142799 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142809 4836 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142816 4836 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142824 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142832 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142840 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142848 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142857 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142865 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142874 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142881 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142890 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142898 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142906 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142914 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142922 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142931 4836 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142939 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142947 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142954 4836 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142962 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142970 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142982 4836 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142990 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.142997 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143005 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143013 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143021 4836 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143029 4836 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143054 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143062 4836 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143070 4836 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143079 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143088 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143096 4836 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143105 4836 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.143835 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.146940 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7" exitCode=255 Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.146991 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7"} Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.148610 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.151613 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.153215 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.158655 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.160249 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.179467 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.190773 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.203101 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.216455 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.220277 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.220378 4836 scope.go:117] "RemoveContainer" containerID="ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.233149 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.244086 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.244112 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.247306 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.249492 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.255358 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.265230 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.266720 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-2f98482294603cfbfeb9c3dfd0700b898d503a9465d75aef235e7f8b53a82e36 WatchSource:0}: Error finding container 2f98482294603cfbfeb9c3dfd0700b898d503a9465d75aef235e7f8b53a82e36: Status 404 returned error can't find the container with id 2f98482294603cfbfeb9c3dfd0700b898d503a9465d75aef235e7f8b53a82e36 Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.271633 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.287963 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.297265 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f36e8c63c29e636673a27a19f008b58ea2e46d730840304dd77a53443d93b3e2 WatchSource:0}: Error finding container f36e8c63c29e636673a27a19f008b58ea2e46d730840304dd77a53443d93b3e2: Status 404 returned error can't find the container with id f36e8c63c29e636673a27a19f008b58ea2e46d730840304dd77a53443d93b3e2 Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.299484 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.304940 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-34dca6ccda2e3bf7b1d17eba62e1f90d6ff7dae2a6eb7f28285218f3fa6f4cb5 WatchSource:0}: Error finding container 34dca6ccda2e3bf7b1d17eba62e1f90d6ff7dae2a6eb7f28285218f3fa6f4cb5: Status 404 returned error can't find the container with id 34dca6ccda2e3bf7b1d17eba62e1f90d6ff7dae2a6eb7f28285218f3fa6f4cb5 Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.475317 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-n8vsx"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.475653 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.475731 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-7wkg5"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.476065 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.477318 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gfrrz"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.477783 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.478532 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.478646 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.478791 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.478898 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.478997 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.482397 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.482833 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.483087 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.483158 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wblv2"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.484580 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.486856 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.486869 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.486885 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.487234 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.487545 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-89st4"] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.487815 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.488065 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492281 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492334 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492503 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492558 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492649 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492712 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492883 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.492954 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.494406 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.497338 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.507381 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.516347 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.525542 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.532990 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.541499 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552583 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552726 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552756 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552776 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552810 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552844 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552864 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552886 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552904 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-multus\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552917 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-hostroot\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552947 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552965 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.552982 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553080 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c414a28-3025-49f7-8213-31f9f2990e97-hosts-file\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553114 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc8qh\" (UniqueName: \"kubernetes.io/projected/8c414a28-3025-49f7-8213-31f9f2990e97-kube-api-access-sc8qh\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553141 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-kubelet\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553159 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-conf-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553176 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-multus-certs\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553196 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553220 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553252 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553280 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-rootfs\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553299 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-bin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553317 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cnibin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553338 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-daemon-config\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553359 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msb79\" (UniqueName: \"kubernetes.io/projected/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-kube-api-access-msb79\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553376 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-system-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553435 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cni-binary-copy\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553456 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-k8s-cni-cncf-io\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553476 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-etc-kubernetes\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553499 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553522 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553547 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553570 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553591 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-proxy-tls\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553607 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-system-cni-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553634 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-os-release\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553652 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-netns\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553679 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553714 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553734 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-socket-dir-parent\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553751 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jt5x\" (UniqueName: \"kubernetes.io/projected/7fcb3e24-f305-4f39-af85-2dc05a0af79f-kube-api-access-2jt5x\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553766 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zbvn\" (UniqueName: \"kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553780 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-os-release\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553807 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r26cl\" (UniqueName: \"kubernetes.io/projected/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-kube-api-access-r26cl\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553839 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553855 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553884 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553901 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553917 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-mcd-auth-proxy-config\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.553935 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cnibin\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.564888 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.575979 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.585009 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.592984 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.604417 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.612898 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.621770 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.630805 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.638312 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.651310 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654428 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654504 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cnibin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654526 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-system-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654548 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-daemon-config\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654584 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msb79\" (UniqueName: \"kubernetes.io/projected/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-kube-api-access-msb79\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654606 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cnibin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.654678 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:17.654632385 +0000 UTC m=+21.014400912 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654728 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-system-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.654855 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.654989 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.655102 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:17.655076066 +0000 UTC m=+21.014844603 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655247 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cni-binary-copy\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655281 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-k8s-cni-cncf-io\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655304 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-etc-kubernetes\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655330 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655382 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-proxy-tls\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655410 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-system-cni-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655437 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655458 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655504 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655533 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655553 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-os-release\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655569 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-netns\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655587 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655607 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655622 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-os-release\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655622 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-system-cni-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655629 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-daemon-config\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655646 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-socket-dir-parent\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655672 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jt5x\" (UniqueName: \"kubernetes.io/projected/7fcb3e24-f305-4f39-af85-2dc05a0af79f-kube-api-access-2jt5x\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655693 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zbvn\" (UniqueName: \"kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655710 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r26cl\" (UniqueName: \"kubernetes.io/projected/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-kube-api-access-r26cl\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655726 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655742 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655749 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-etc-kubernetes\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655761 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655812 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-mcd-auth-proxy-config\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655832 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655850 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661440 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fcb3e24-f305-4f39-af85-2dc05a0af79f-cni-binary-copy\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.656156 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cnibin\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661542 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661564 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661593 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661620 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661642 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661659 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661683 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661706 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-multus\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661727 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-hostroot\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661751 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661771 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661801 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661825 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661852 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c414a28-3025-49f7-8213-31f9f2990e97-hosts-file\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661874 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc8qh\" (UniqueName: \"kubernetes.io/projected/8c414a28-3025-49f7-8213-31f9f2990e97-kube-api-access-sc8qh\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661893 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-multus-certs\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661915 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661943 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-kubelet\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661970 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-conf-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.661991 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-bin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662018 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662057 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662082 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-rootfs\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662159 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-rootfs\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662206 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662345 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662420 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:17.662398788 +0000 UTC m=+21.022167315 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662449 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.655706 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-k8s-cni-cncf-io\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662492 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662516 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662552 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662562 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662603 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-netns\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662623 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662550 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662571 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662697 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-os-release\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.662738 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:17.662710306 +0000 UTC m=+21.022478823 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662768 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-socket-dir-parent\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662779 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-multus\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662817 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-hostroot\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662818 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-cni-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662851 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662881 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662893 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.663001 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.663024 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.663062 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: E0122 09:42:16.663110 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:17.663097376 +0000 UTC m=+21.022865893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663135 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663154 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663213 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662531 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-os-release\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.662524 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663275 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8c414a28-3025-49f7-8213-31f9f2990e97-hosts-file\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663305 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663318 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663344 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-run-multus-certs\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663377 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-mcd-auth-proxy-config\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663401 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663428 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663540 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-cni-bin\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663564 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-host-var-lib-kubelet\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663588 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fcb3e24-f305-4f39-af85-2dc05a0af79f-multus-conf-dir\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663613 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663635 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cnibin\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.663807 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.664164 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.664448 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-proxy-tls\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.664472 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-cni-binary-copy\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.668464 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.676066 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.678459 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msb79\" (UniqueName: \"kubernetes.io/projected/17ea94f0-dd58-4f5b-b61c-b59c074af1f3-kube-api-access-msb79\") pod \"machine-config-daemon-n8vsx\" (UID: \"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\") " pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.679218 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zbvn\" (UniqueName: \"kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn\") pod \"ovnkube-node-wblv2\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.680133 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc8qh\" (UniqueName: \"kubernetes.io/projected/8c414a28-3025-49f7-8213-31f9f2990e97-kube-api-access-sc8qh\") pod \"node-resolver-7wkg5\" (UID: \"8c414a28-3025-49f7-8213-31f9f2990e97\") " pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.682406 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r26cl\" (UniqueName: \"kubernetes.io/projected/2aa3f461-19ed-4700-94b6-8c8dd71c6a21-kube-api-access-r26cl\") pod \"multus-additional-cni-plugins-gfrrz\" (UID: \"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\") " pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.685118 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jt5x\" (UniqueName: \"kubernetes.io/projected/7fcb3e24-f305-4f39-af85-2dc05a0af79f-kube-api-access-2jt5x\") pod \"multus-89st4\" (UID: \"7fcb3e24-f305-4f39-af85-2dc05a0af79f\") " pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.693063 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:16Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.715546 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:16Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.793153 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.803906 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-7wkg5" Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.805030 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17ea94f0_dd58_4f5b_b61c_b59c074af1f3.slice/crio-f5c535b682398a9ecba48bcb6c195c42b1bb98b9dc3120ecefeeda4a912298b9 WatchSource:0}: Error finding container f5c535b682398a9ecba48bcb6c195c42b1bb98b9dc3120ecefeeda4a912298b9: Status 404 returned error can't find the container with id f5c535b682398a9ecba48bcb6c195c42b1bb98b9dc3120ecefeeda4a912298b9 Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.810659 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.819634 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-89st4" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.822675 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.824183 4836 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824347 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824740 4836 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824775 4836 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824803 4836 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824830 4836 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824856 4836 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824882 4836 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824908 4836 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824931 4836 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.824956 4836 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825118 4836 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825181 4836 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825176 4836 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825203 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825216 4836 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825230 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825232 4836 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825240 4836 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825261 4836 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825271 4836 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825289 4836 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825300 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825342 4836 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825361 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825379 4836 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825400 4836 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825420 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825442 4836 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825458 4836 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825470 4836 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.825491 4836 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 22 09:42:16 crc kubenswrapper[4836]: W0122 09:42:16.835124 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aa3f461_19ed_4700_94b6_8c8dd71c6a21.slice/crio-0f4a16843705f6badd464f7e267f94b8edac768b77d2e4175412d5f1f082202d WatchSource:0}: Error finding container 0f4a16843705f6badd464f7e267f94b8edac768b77d2e4175412d5f1f082202d: Status 404 returned error can't find the container with id 0f4a16843705f6badd464f7e267f94b8edac768b77d2e4175412d5f1f082202d Jan 22 09:42:16 crc kubenswrapper[4836]: I0122 09:42:16.942163 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 23:05:57.138501186 +0000 UTC Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.020960 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.021486 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.023217 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.024018 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.025005 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.025532 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.026114 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.030318 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.030888 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.031501 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.032573 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.033114 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.034182 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.034726 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.035263 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.043590 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.044164 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.045312 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.045715 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.046279 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.053576 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.054165 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.055181 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.055627 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.056656 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.058233 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.058311 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.058823 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.067399 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.067915 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.068491 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.069460 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.070213 4836 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.070364 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.074632 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.076706 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.077293 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.080693 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.085122 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.085745 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.098255 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.099265 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.099863 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.100873 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.101577 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.102175 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.102687 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.103240 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.111761 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.112309 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.113604 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.114090 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.119999 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.120593 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.130315 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.130456 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.131284 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.131782 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.155342 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.161354 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerStarted","Data":"0f4a16843705f6badd464f7e267f94b8edac768b77d2e4175412d5f1f082202d"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.162722 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.162770 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f36e8c63c29e636673a27a19f008b58ea2e46d730840304dd77a53443d93b3e2"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.175822 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97" exitCode=0 Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.175953 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.175984 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"be5fbf6a39c5b074c978ee3d390727c7bc9d8187b20407412a81a3e8f12974a8"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.185214 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.192237 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.192280 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.192291 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2f98482294603cfbfeb9c3dfd0700b898d503a9465d75aef235e7f8b53a82e36"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.206447 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7wkg5" event={"ID":"8c414a28-3025-49f7-8213-31f9f2990e97","Type":"ContainerStarted","Data":"28e83bf8d1f2d62e4417c2b281a2e2261141d6ec5d5c514637b5f5beba095ebf"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.228866 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.233518 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.233560 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.233569 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"f5c535b682398a9ecba48bcb6c195c42b1bb98b9dc3120ecefeeda4a912298b9"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.236913 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.238156 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.238916 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.247470 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerStarted","Data":"07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.247516 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerStarted","Data":"1433eef4f10a8a0a4f086ad658aea51a0703d2f4ee987b506c3d9a25a8273d73"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.253152 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.257555 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"34dca6ccda2e3bf7b1d17eba62e1f90d6ff7dae2a6eb7f28285218f3fa6f4cb5"} Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.269674 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.295499 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.307332 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.323839 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.334296 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.344029 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.355623 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.370522 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.382024 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.396492 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.411960 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.427506 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.440673 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.465295 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.478660 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.505244 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.670411 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.671618 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.671714 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.671773 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:19.671754731 +0000 UTC m=+23.031523248 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.671825 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.671839 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.671849 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.671890 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:19.671875404 +0000 UTC m=+23.031643921 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.671877 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.671936 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.671974 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672035 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672075 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672102 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:19.67209266 +0000 UTC m=+23.031861177 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672108 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672125 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672117 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:19.67211184 +0000 UTC m=+23.031880487 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672136 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.672164 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:19.672154611 +0000 UTC m=+23.031923128 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.744760 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.748846 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.787464 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.788311 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.800502 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.806787 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.808398 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.819948 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.824169 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.864886 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.915275 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.928566 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.942964 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 21:06:56.763588036 +0000 UTC Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.989545 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.996510 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.996567 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.996631 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:17 crc kubenswrapper[4836]: I0122 09:42:17.996569 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.996783 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:17 crc kubenswrapper[4836]: E0122 09:42:17.996896 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.028510 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.053259 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.068241 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.091804 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.106777 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.116379 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.175916 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.176361 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.219237 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.245484 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.247484 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.262576 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d" exitCode=0 Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.262642 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.264975 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-7wkg5" event={"ID":"8c414a28-3025-49f7-8213-31f9f2990e97","Type":"ContainerStarted","Data":"99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.268881 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.269066 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.269160 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.269235 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.269295 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.269355 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576"} Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.270789 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.275258 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.279922 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.291201 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.291331 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.303590 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.310090 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.317223 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.329539 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.344004 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.345404 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.356743 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.375400 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.391862 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.406125 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.410284 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.428935 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.441193 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:18Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:18 crc kubenswrapper[4836]: I0122 09:42:18.943323 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 18:58:47.387880659 +0000 UTC Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.103122 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8zwn4"] Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.103528 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.105638 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.106794 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.107399 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.107687 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.119571 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.135557 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.147585 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.161582 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.183340 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.187945 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d7820ac-f475-497d-ac8c-623ef565faee-serviceca\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.188010 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d7820ac-f475-497d-ac8c-623ef565faee-host\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.188095 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqddz\" (UniqueName: \"kubernetes.io/projected/0d7820ac-f475-497d-ac8c-623ef565faee-kube-api-access-rqddz\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.201136 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.212440 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.229872 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.239613 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.251448 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.273410 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a"} Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.274654 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.275681 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11" exitCode=0 Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.275723 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11"} Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.289408 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqddz\" (UniqueName: \"kubernetes.io/projected/0d7820ac-f475-497d-ac8c-623ef565faee-kube-api-access-rqddz\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.289473 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d7820ac-f475-497d-ac8c-623ef565faee-serviceca\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.289512 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d7820ac-f475-497d-ac8c-623ef565faee-host\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.289589 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d7820ac-f475-497d-ac8c-623ef565faee-host\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.290762 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d7820ac-f475-497d-ac8c-623ef565faee-serviceca\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.293380 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.316669 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqddz\" (UniqueName: \"kubernetes.io/projected/0d7820ac-f475-497d-ac8c-623ef565faee-kube-api-access-rqddz\") pod \"node-ca-8zwn4\" (UID: \"0d7820ac-f475-497d-ac8c-623ef565faee\") " pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.322985 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.338229 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.356415 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.372705 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.382742 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.397560 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.416938 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.419200 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8zwn4" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.429896 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.440801 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: W0122 09:42:19.441821 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d7820ac_f475_497d_ac8c_623ef565faee.slice/crio-a1bdb7d643a01c063634b87fcfadd3e04d3ec9bcdb3bf8bf3e03fb567b19f422 WatchSource:0}: Error finding container a1bdb7d643a01c063634b87fcfadd3e04d3ec9bcdb3bf8bf3e03fb567b19f422: Status 404 returned error can't find the container with id a1bdb7d643a01c063634b87fcfadd3e04d3ec9bcdb3bf8bf3e03fb567b19f422 Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.452496 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.468632 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.485858 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.504315 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.522169 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.694415 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.694537 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694585 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:23.694548704 +0000 UTC m=+27.054317261 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694639 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.694643 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694686 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:23.694673368 +0000 UTC m=+27.054441875 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.694736 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.694788 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694856 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694878 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694908 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:23.694884983 +0000 UTC m=+27.054653500 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694910 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694924 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694970 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694987 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:23.694971005 +0000 UTC m=+27.054739522 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.694995 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.695016 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.695116 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:23.695097918 +0000 UTC m=+27.054866475 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.787821 4836 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.790881 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.790937 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.790951 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.791105 4836 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.799325 4836 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.799520 4836 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.800502 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.800547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.800564 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.800583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.800596 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.813185 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.816887 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.816929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.816938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.816953 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.816963 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.837660 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.841096 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.841153 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.841170 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.841195 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.841213 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.854774 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.858228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.858259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.858267 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.858281 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.858291 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.870634 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.874050 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.874093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.874119 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.874134 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.874143 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.885144 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:19Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.885257 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.886593 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.886635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.886644 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.886659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.886669 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.944434 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 17:52:31.537053113 +0000 UTC Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.989824 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.989875 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.989888 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.989908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.989921 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:19Z","lastTransitionTime":"2026-01-22T09:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.997085 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.997170 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.997207 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:19 crc kubenswrapper[4836]: I0122 09:42:19.997086 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.997294 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:19 crc kubenswrapper[4836]: E0122 09:42:19.997436 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.091848 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.091891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.091902 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.091919 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.091930 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.194353 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.194398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.194413 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.194433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.194447 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.283158 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8zwn4" event={"ID":"0d7820ac-f475-497d-ac8c-623ef565faee","Type":"ContainerStarted","Data":"ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.283240 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8zwn4" event={"ID":"0d7820ac-f475-497d-ac8c-623ef565faee","Type":"ContainerStarted","Data":"a1bdb7d643a01c063634b87fcfadd3e04d3ec9bcdb3bf8bf3e03fb567b19f422"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.286239 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf" exitCode=0 Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.286286 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.297601 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.297639 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.297651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.297667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.297679 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.304679 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.321072 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.346071 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.366209 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.399437 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.399475 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.399485 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.399500 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.399512 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.402084 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.414522 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.424550 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.439391 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.452963 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.464269 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.481893 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.494340 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.501954 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.501985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.501996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.502013 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.502024 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.508659 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.522269 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.534514 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.550527 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.562494 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.575304 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.587571 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.600080 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.604186 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.604242 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.604259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.604287 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.604305 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.613708 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.625778 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.643923 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.659405 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.675749 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.688776 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:20Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.707301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.707360 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.707376 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.707399 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.707419 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.809858 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.809898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.809911 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.809934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.809946 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.912851 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.912895 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.912907 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.912926 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.912939 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:20Z","lastTransitionTime":"2026-01-22T09:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:20 crc kubenswrapper[4836]: I0122 09:42:20.945309 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 12:26:05.502530553 +0000 UTC Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.015791 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.015858 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.015939 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.015975 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.016002 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.119508 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.119566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.119607 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.119625 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.119637 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.223610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.223676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.223695 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.223722 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.223741 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.297931 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.303222 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548" exitCode=0 Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.303286 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.319635 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.331343 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.331426 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.331454 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.331489 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.331508 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.347833 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.371437 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.403350 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.427618 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.435367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.435474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.435517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.435549 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.435587 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.447284 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.468450 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.479512 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.497152 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.511194 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.524136 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.536190 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.537278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.537312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.537321 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.537335 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.537345 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.547686 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.640081 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.640122 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.640130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.640144 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.640153 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.743078 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.743120 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.743132 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.743149 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.743161 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.747160 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.750952 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.759699 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.779896 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.799445 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.819532 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.838155 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.845658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.845735 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.845761 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.845794 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.845811 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.852810 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.868113 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.884471 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.896122 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.908713 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.921538 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.941111 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.946518 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 23:35:24.958768434 +0000 UTC Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.949020 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.949081 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.949093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.949112 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.949123 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:21Z","lastTransitionTime":"2026-01-22T09:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.954318 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.972510 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.989896 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.996896 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.996921 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:21 crc kubenswrapper[4836]: I0122 09:42:21.996945 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:21 crc kubenswrapper[4836]: E0122 09:42:21.997022 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:21 crc kubenswrapper[4836]: E0122 09:42:21.997183 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:21 crc kubenswrapper[4836]: E0122 09:42:21.997276 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.013350 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.036396 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.052178 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.052231 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.052247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.052270 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.052286 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.066662 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.086144 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.106719 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.122417 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.140336 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.155164 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.155248 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.155293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.155342 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.155369 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.156358 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.175314 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.188698 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.205462 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.218334 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.234516 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.258549 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.258590 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.258613 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.258630 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.258644 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.310155 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db" exitCode=0 Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.310271 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.332455 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.361410 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.361473 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.361491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.361522 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.361541 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.365615 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.383420 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.413808 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.436554 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.463587 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.464992 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.465028 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.465064 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.465087 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.465099 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.490245 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.508350 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.520959 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.535462 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.547605 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.561091 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.567243 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.567284 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.567295 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.567311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.567322 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.571719 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.580770 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:22Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.670700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.670743 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.670752 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.670768 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.670778 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.774615 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.774720 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.774748 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.774781 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.774804 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.877365 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.877431 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.877448 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.877472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.877485 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.947436 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 03:39:35.686787001 +0000 UTC Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.980570 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.980650 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.980674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.980721 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:22 crc kubenswrapper[4836]: I0122 09:42:22.980756 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:22Z","lastTransitionTime":"2026-01-22T09:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.082830 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.082878 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.082889 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.082906 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.082918 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.185531 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.185629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.185654 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.185685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.185713 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.288363 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.288433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.288453 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.288477 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.288553 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.319738 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.320091 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.325910 4836 generic.go:334] "Generic (PLEG): container finished" podID="2aa3f461-19ed-4700-94b6-8c8dd71c6a21" containerID="8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911" exitCode=0 Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.325962 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerDied","Data":"8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.336991 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.353401 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.354338 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.368828 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.383765 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.390868 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.390930 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.390948 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.390971 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.390988 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.403361 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.415497 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.431539 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.447085 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.461273 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.482619 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.494630 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.494675 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.494686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.494704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.494715 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.497288 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.513660 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.526784 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.541462 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.557527 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.570503 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.587158 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.597517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.597561 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.597572 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.597591 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.597614 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.598541 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.611293 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.626657 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.640724 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.661634 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.675593 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.688717 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.699608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.699658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.699672 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.699691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.699703 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.703021 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.715353 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.726512 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.730752 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.730945 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.730958 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.730926411 +0000 UTC m=+35.090694938 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731115 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.731125 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731171 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.731157327 +0000 UTC m=+35.090925844 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.731195 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.731256 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731351 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731442 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.731422003 +0000 UTC m=+35.091190560 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731452 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731483 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731485 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731503 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731525 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731546 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731594 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.731573057 +0000 UTC m=+35.091341604 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.731626 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.731610618 +0000 UTC m=+35.091379175 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.742649 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:23Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.803257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.803340 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.803367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.803398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.803421 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.906438 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.906497 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.906514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.906537 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.906553 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:23Z","lastTransitionTime":"2026-01-22T09:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.947737 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 16:52:53.118305284 +0000 UTC Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.996124 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.996226 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:23 crc kubenswrapper[4836]: I0122 09:42:23.996329 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.996329 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.996588 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:23 crc kubenswrapper[4836]: E0122 09:42:23.996679 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.010097 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.010184 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.010208 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.010230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.010247 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.113407 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.113501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.113522 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.113583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.113602 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.216263 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.216360 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.216384 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.216411 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.216432 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.318914 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.318968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.318980 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.318995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.319008 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.332146 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" event={"ID":"2aa3f461-19ed-4700-94b6-8c8dd71c6a21","Type":"ContainerStarted","Data":"eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.332222 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.333511 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.352643 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.358199 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.370323 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.384484 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.394346 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.409999 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.421130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.421177 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.421188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.421204 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.421214 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.424762 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.435966 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.448102 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.458497 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.472334 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.487527 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.499858 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.521962 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.523772 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.523813 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.523823 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.523838 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.523846 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.536969 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.553941 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.576124 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.591086 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.621397 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.625869 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.625908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.625920 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.625939 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.625952 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.636534 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.661841 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.684403 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.703281 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.721120 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.728397 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.728438 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.728450 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.728466 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.728479 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.734295 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.747013 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.758670 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.770907 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.780847 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:24Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.830753 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.830799 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.830810 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.830823 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.830832 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.933190 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.933257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.933267 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.933341 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.933351 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:24Z","lastTransitionTime":"2026-01-22T09:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:24 crc kubenswrapper[4836]: I0122 09:42:24.947917 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 04:33:22.354471029 +0000 UTC Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.035414 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.035462 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.035478 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.035502 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.035520 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.141784 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.141834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.141850 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.141869 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.141881 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.244622 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.244664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.244680 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.244701 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.244718 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.335019 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.347387 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.347429 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.347444 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.347464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.347480 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.449991 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.450062 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.450074 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.450088 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.450099 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.553104 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.553166 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.553183 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.553207 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.553228 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.655599 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.655646 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.655655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.655670 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.655680 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.757622 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.757673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.757686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.757705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.757719 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.860210 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.860279 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.860302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.860331 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.860353 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.948811 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 00:12:28.4547047 +0000 UTC Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.963808 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.963864 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.963882 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.963906 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.963923 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:25Z","lastTransitionTime":"2026-01-22T09:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.996463 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.996556 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:25 crc kubenswrapper[4836]: I0122 09:42:25.996489 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:25 crc kubenswrapper[4836]: E0122 09:42:25.996652 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:25 crc kubenswrapper[4836]: E0122 09:42:25.996842 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:25 crc kubenswrapper[4836]: E0122 09:42:25.997059 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.066926 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.066971 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.066986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.067006 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.067022 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.170404 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.170487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.170511 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.170544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.170568 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.273328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.273574 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.273686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.273766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.273859 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.340872 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/0.log" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.343988 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86" exitCode=1 Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.344020 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.345132 4836 scope.go:117] "RemoveContainer" containerID="24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.376663 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.376704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.376715 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.376734 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.376749 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.379354 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:25Z\\\",\\\"message\\\":\\\"I0122 09:42:25.326601 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 09:42:25.326618 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:42:25.326625 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:42:25.326650 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 09:42:25.326660 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:42:25.326690 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:42:25.326701 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:42:25.326698 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0122 09:42:25.326721 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 09:42:25.326702 6077 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 09:42:25.326739 6077 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 09:42:25.326757 6077 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 09:42:25.326795 6077 factory.go:656] Stopping watch factory\\\\nI0122 09:42:25.326812 6077 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.403006 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.423450 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.449018 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.469601 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.479563 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.479614 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.479632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.479659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.479677 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.482845 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.495921 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.517008 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.533197 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.552984 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.568439 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.584585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.584612 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.584622 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.584638 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.584649 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.588248 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.606361 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.629144 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:26Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.687501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.687539 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.687546 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.687561 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.687572 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.790621 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.790747 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.790765 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.790791 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.790814 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.894514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.894611 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.894634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.894659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.894675 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.949412 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 21:12:40.868648775 +0000 UTC Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.998474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.998550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.998561 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.998586 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:26 crc kubenswrapper[4836]: I0122 09:42:26.998604 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:26Z","lastTransitionTime":"2026-01-22T09:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.012929 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.031174 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.052724 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.077634 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.100896 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.100936 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.100950 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.100970 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.100984 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.101364 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.120588 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.138661 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.168967 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.190278 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.203507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.203576 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.203592 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.203610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.203623 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.208242 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.233912 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:25Z\\\",\\\"message\\\":\\\"I0122 09:42:25.326601 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 09:42:25.326618 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:42:25.326625 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:42:25.326650 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 09:42:25.326660 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:42:25.326690 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:42:25.326701 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:42:25.326698 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0122 09:42:25.326721 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 09:42:25.326702 6077 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 09:42:25.326739 6077 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 09:42:25.326757 6077 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 09:42:25.326795 6077 factory.go:656] Stopping watch factory\\\\nI0122 09:42:25.326812 6077 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.249935 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.268772 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.285980 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.306031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.306103 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.306118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.306139 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.306153 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.351602 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/0.log" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.356119 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.356405 4836 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.387025 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.407081 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.409472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.409517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.409530 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.409549 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.409562 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.428430 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.448426 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.470222 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.491248 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.503466 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.512324 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.512363 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.512374 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.512392 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.512403 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.514585 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.526546 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.540156 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.553251 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.574696 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:25Z\\\",\\\"message\\\":\\\"I0122 09:42:25.326601 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 09:42:25.326618 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:42:25.326625 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:42:25.326650 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 09:42:25.326660 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:42:25.326690 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:42:25.326701 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:42:25.326698 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0122 09:42:25.326721 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 09:42:25.326702 6077 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 09:42:25.326739 6077 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 09:42:25.326757 6077 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 09:42:25.326795 6077 factory.go:656] Stopping watch factory\\\\nI0122 09:42:25.326812 6077 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.591539 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.606957 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.614671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.614710 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.614727 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.614749 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.614766 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.717272 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.717308 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.717319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.717335 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.717346 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.819651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.819697 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.819713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.819734 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.819749 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.922200 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.922239 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.922248 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.922266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.922275 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:27Z","lastTransitionTime":"2026-01-22T09:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.949880 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 23:39:48.793132302 +0000 UTC Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.996236 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.996275 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:27 crc kubenswrapper[4836]: I0122 09:42:27.996305 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:27 crc kubenswrapper[4836]: E0122 09:42:27.996364 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:27 crc kubenswrapper[4836]: E0122 09:42:27.996434 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:27 crc kubenswrapper[4836]: E0122 09:42:27.996508 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.024291 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.024348 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.024365 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.024389 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.024408 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.127945 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.128032 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.128098 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.128125 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.128173 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.231104 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.231200 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.231220 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.231247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.231299 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.334524 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.334577 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.334588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.334607 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.334618 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.361909 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/1.log" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.362958 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/0.log" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.367565 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a" exitCode=1 Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.367612 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.367667 4836 scope.go:117] "RemoveContainer" containerID="24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.368483 4836 scope.go:117] "RemoveContainer" containerID="35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a" Jan 22 09:42:28 crc kubenswrapper[4836]: E0122 09:42:28.368675 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.390431 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.409612 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.428902 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.437293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.437353 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.437411 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.437437 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.437456 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.445581 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.463300 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.489287 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.513426 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.530685 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.540180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.540247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.540264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.540284 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.540298 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.548341 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.564993 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.583978 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.606268 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.626304 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.643021 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.643296 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.643445 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.643579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.643746 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.654685 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:25Z\\\",\\\"message\\\":\\\"I0122 09:42:25.326601 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 09:42:25.326618 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:42:25.326625 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:42:25.326650 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 09:42:25.326660 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:42:25.326690 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:42:25.326701 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:42:25.326698 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0122 09:42:25.326721 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 09:42:25.326702 6077 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 09:42:25.326739 6077 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 09:42:25.326757 6077 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 09:42:25.326795 6077 factory.go:656] Stopping watch factory\\\\nI0122 09:42:25.326812 6077 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.746483 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.746552 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.746565 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.746583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.746596 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.849516 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.849595 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.849613 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.849638 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.849655 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.919852 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8"] Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.920667 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.925446 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.925592 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.943429 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.950660 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 03:54:40.955415223 +0000 UTC Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.952845 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.952914 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.952940 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.953007 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.953033 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:28Z","lastTransitionTime":"2026-01-22T09:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.959296 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.978606 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.990360 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.990406 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfb9h\" (UniqueName: \"kubernetes.io/projected/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-kube-api-access-jfb9h\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.990448 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.990499 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:28 crc kubenswrapper[4836]: I0122 09:42:28.998847 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:28Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.015386 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.046783 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.055838 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.055882 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.055893 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.055910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.055924 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.058392 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.069694 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.079851 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.089729 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.091129 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.091197 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.091232 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfb9h\" (UniqueName: \"kubernetes.io/projected/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-kube-api-access-jfb9h\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.091282 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.091904 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.092230 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.098482 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.103516 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.109747 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfb9h\" (UniqueName: \"kubernetes.io/projected/09bc4a6f-04fc-4e7a-bd92-ceac6f457de8-kube-api-access-jfb9h\") pod \"ovnkube-control-plane-749d76644c-6qnv8\" (UID: \"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.123269 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.138302 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.158490 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.158542 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.158553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.158571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.158586 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.159927 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24cd07dab4a1ca2623efe097fe85a036e2af44b3aeffb463a9d83c9a0c45ed86\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:25Z\\\",\\\"message\\\":\\\"I0122 09:42:25.326601 6077 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0122 09:42:25.326618 6077 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:42:25.326625 6077 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:42:25.326650 6077 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0122 09:42:25.326660 6077 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:42:25.326690 6077 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:42:25.326701 6077 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:42:25.326698 6077 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:42:25.326666 6077 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0122 09:42:25.326721 6077 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0122 09:42:25.326702 6077 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0122 09:42:25.326739 6077 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0122 09:42:25.326757 6077 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0122 09:42:25.326795 6077 factory.go:656] Stopping watch factory\\\\nI0122 09:42:25.326812 6077 ovnkube.go:599] Stopped ovnkube\\\\nI0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.174173 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.241788 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.261566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.261615 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.261633 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.261656 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.261673 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.363790 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.363837 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.363849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.363867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.363882 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.377250 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" event={"ID":"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8","Type":"ContainerStarted","Data":"aba8e712a67afc3cd4f16da27a26befa93061dae7783b937bbb281f0c12b6ad0"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.382230 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/1.log" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.386254 4836 scope.go:117] "RemoveContainer" containerID="35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a" Jan 22 09:42:29 crc kubenswrapper[4836]: E0122 09:42:29.386424 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.402640 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.417648 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.432576 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.445551 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.462627 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.467034 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.467093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.467101 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.467118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.467127 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.477808 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.490195 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.506926 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.528019 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.544813 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.570973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.571031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.571058 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.571078 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.571090 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.573552 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.591493 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.612860 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.630970 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.649825 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:29Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.674346 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.674424 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.674442 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.674469 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.674545 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.778423 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.778494 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.778514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.778538 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.778557 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.881128 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.881167 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.881179 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.881199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.881209 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.951054 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 03:27:57.903840911 +0000 UTC Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.984364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.984405 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.984417 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.984435 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.984448 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:29Z","lastTransitionTime":"2026-01-22T09:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.996757 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.996762 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:29 crc kubenswrapper[4836]: E0122 09:42:29.996938 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:29 crc kubenswrapper[4836]: I0122 09:42:29.996785 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:29 crc kubenswrapper[4836]: E0122 09:42:29.997058 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:29 crc kubenswrapper[4836]: E0122 09:42:29.997167 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.087908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.087949 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.087959 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.087976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.087991 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.191498 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.191533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.191544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.191560 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.191571 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.238202 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.238263 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.238281 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.238310 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.238329 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.260329 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.265710 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.265767 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.265785 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.265810 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.265830 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.290757 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.296613 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.296655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.296667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.296685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.296697 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.319396 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.327451 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.327528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.327550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.327578 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.327598 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.349156 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.354442 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.354511 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.354544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.354577 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.354597 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.374281 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.374572 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.376887 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.376944 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.376962 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.376990 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.377009 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.392452 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" event={"ID":"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8","Type":"ContainerStarted","Data":"f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.392508 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" event={"ID":"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8","Type":"ContainerStarted","Data":"ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.413902 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.416229 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6kkmw"] Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.416871 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.416978 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.431264 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.445813 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.463963 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.479490 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.479550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.479569 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.479594 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.479612 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.481091 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.495134 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.508299 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.508406 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jr6q\" (UniqueName: \"kubernetes.io/projected/9ab725c3-0744-400a-b8d7-66e1165476e5-kube-api-access-8jr6q\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.509194 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.523328 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.538506 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.555646 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.569811 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.582578 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.582623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.582639 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.582661 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.582676 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.593326 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.609004 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.609051 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jr6q\" (UniqueName: \"kubernetes.io/projected/9ab725c3-0744-400a-b8d7-66e1165476e5-kube-api-access-8jr6q\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.609295 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:30 crc kubenswrapper[4836]: E0122 09:42:30.609420 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:31.109393379 +0000 UTC m=+34.469161936 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.610075 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.628987 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.631681 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jr6q\" (UniqueName: \"kubernetes.io/projected/9ab725c3-0744-400a-b8d7-66e1165476e5-kube-api-access-8jr6q\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.650220 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.664651 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.681898 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.686472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.686507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.686519 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.686536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.686549 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.701292 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.732930 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.749786 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.767487 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.782146 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.789385 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.789424 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.789433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.789450 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.789461 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.802771 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.820107 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.833968 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.855575 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.872357 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.890378 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.892188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.892223 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.892234 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.892251 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.892263 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.907657 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.925718 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.941779 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:30Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.952077 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 14:22:57.675519827 +0000 UTC Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.995459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.995509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.995525 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.995544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:30 crc kubenswrapper[4836]: I0122 09:42:30.995558 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:30Z","lastTransitionTime":"2026-01-22T09:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.097553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.097585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.097593 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.097605 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.097614 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.115245 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.115395 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.115444 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:32.115430276 +0000 UTC m=+35.475198793 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.200391 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.200520 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.200568 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.200606 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.200630 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.303627 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.303700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.303719 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.303747 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.303766 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.407525 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.407609 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.407626 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.407669 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.407683 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.510585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.510671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.510695 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.510724 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.510745 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.613997 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.614114 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.614139 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.614169 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.614186 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.717361 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.717463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.717501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.717538 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.717562 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.820207 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.820309 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.820333 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.820724 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.820775 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.822862 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823029 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:42:47.823002987 +0000 UTC m=+51.182771544 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.823183 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.823278 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.823345 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823383 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.823411 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823463 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:47.823435899 +0000 UTC m=+51.183204476 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823551 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823570 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823609 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823609 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823629 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823640 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823662 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823641 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:47.823616204 +0000 UTC m=+51.183384771 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823836 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:47.823763858 +0000 UTC m=+51.183532415 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.823883 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:47.82386828 +0000 UTC m=+51.183636837 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.923841 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.923910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.923922 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.923942 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.923976 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:31Z","lastTransitionTime":"2026-01-22T09:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.952561 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 15:30:27.458314431 +0000 UTC Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.997081 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.997168 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.997184 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:31 crc kubenswrapper[4836]: I0122 09:42:31.997077 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.997320 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.997449 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.997609 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:31 crc kubenswrapper[4836]: E0122 09:42:31.997704 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.026784 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.026849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.026870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.026896 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.026915 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.126704 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:32 crc kubenswrapper[4836]: E0122 09:42:32.126929 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:32 crc kubenswrapper[4836]: E0122 09:42:32.127093 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:34.127023008 +0000 UTC m=+37.486791555 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.129623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.129694 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.129712 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.129741 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.129760 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.232356 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.232417 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.232434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.232457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.232475 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.334656 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.334697 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.334705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.334718 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.334728 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.437639 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.437700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.437717 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.437744 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.437762 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.540488 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.540577 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.540605 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.540639 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.540660 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.642942 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.642986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.643007 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.643027 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.643072 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.746367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.746416 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.746436 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.746455 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.746466 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.848804 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.848854 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.848870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.848897 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.848923 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.951847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.951913 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.951936 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.951968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.951994 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:32Z","lastTransitionTime":"2026-01-22T09:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:32 crc kubenswrapper[4836]: I0122 09:42:32.953232 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:56:41.372708451 +0000 UTC Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.055608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.055669 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.055685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.055707 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.055724 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.158898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.158972 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.158999 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.159032 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.159159 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.262446 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.262510 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.262526 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.262555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.262573 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.367216 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.367256 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.367271 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.367360 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.367379 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.470436 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.470474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.470486 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.470501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.470512 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.574085 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.574157 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.574175 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.574199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.574216 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.653232 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.676714 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.676808 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.676834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.676866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.676888 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.678384 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.699954 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.716004 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.732457 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.754424 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.770544 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.779472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.779523 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.779536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.779552 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.779564 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.790036 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.809685 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.826188 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.840462 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.863829 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.882477 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.882536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.882558 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.882580 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.882594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.892583 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.910940 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.939491 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.954240 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 11:59:24.285239713 +0000 UTC Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.955109 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.974297 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:33Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.985685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.985749 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.985767 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.985798 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.985820 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:33Z","lastTransitionTime":"2026-01-22T09:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.997062 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.997159 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.997198 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:33 crc kubenswrapper[4836]: E0122 09:42:33.997339 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:33 crc kubenswrapper[4836]: I0122 09:42:33.997391 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:33 crc kubenswrapper[4836]: E0122 09:42:33.997518 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:33 crc kubenswrapper[4836]: E0122 09:42:33.997625 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:33 crc kubenswrapper[4836]: E0122 09:42:33.997766 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.089677 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.089745 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.089768 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.089799 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.089821 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.148343 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:34 crc kubenswrapper[4836]: E0122 09:42:34.148607 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:34 crc kubenswrapper[4836]: E0122 09:42:34.148716 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:38.148681152 +0000 UTC m=+41.508449709 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.193760 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.193820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.193837 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.193860 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.193878 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.296980 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.297124 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.297174 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.297201 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.297218 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.401385 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.401450 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.401468 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.401498 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.401516 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.504251 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.504292 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.504305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.504323 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.504337 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.607187 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.607243 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.607257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.607277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.607290 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.710143 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.710180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.710188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.710201 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.710209 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.813195 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.813273 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.813291 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.813317 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.813335 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.915831 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.916083 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.916102 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.916128 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.916148 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:34Z","lastTransitionTime":"2026-01-22T09:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:34 crc kubenswrapper[4836]: I0122 09:42:34.954742 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 22:07:05.126385434 +0000 UTC Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.018488 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.018549 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.018566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.018589 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.018607 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.121746 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.121792 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.121808 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.121830 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.121847 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.224775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.224871 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.224895 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.224929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.224953 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.327771 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.327843 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.327861 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.327885 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.327905 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.431000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.431087 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.431100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.431118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.431129 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.533536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.533628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.533659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.533687 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.533704 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.637355 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.637432 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.637443 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.637464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.637477 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.741158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.741231 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.741256 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.741285 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.741306 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.843926 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.843976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.843993 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.844010 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.844022 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.947018 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.947110 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.947127 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.947152 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.947171 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:35Z","lastTransitionTime":"2026-01-22T09:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.954973 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 02:18:12.133819773 +0000 UTC Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.996862 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.996862 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.996897 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:35 crc kubenswrapper[4836]: I0122 09:42:35.997032 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:35 crc kubenswrapper[4836]: E0122 09:42:35.997234 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:35 crc kubenswrapper[4836]: E0122 09:42:35.997641 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:35 crc kubenswrapper[4836]: E0122 09:42:35.997778 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:35 crc kubenswrapper[4836]: E0122 09:42:35.997782 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.050308 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.050382 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.050429 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.050464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.050486 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.154077 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.154135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.154154 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.154179 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.154196 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.257259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.257319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.257336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.257361 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.257378 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.360733 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.360832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.360852 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.360878 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.360897 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.462637 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.462681 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.462694 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.462711 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.462724 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.565114 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.565230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.565247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.565268 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.565294 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.667140 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.667183 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.667194 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.667209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.667218 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.769754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.769798 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.769815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.769838 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.769855 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.872752 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.872833 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.872847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.872868 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.872883 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.955853 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 02:06:09.105684622 +0000 UTC Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.975642 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.975708 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.975725 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.975753 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:36 crc kubenswrapper[4836]: I0122 09:42:36.975775 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:36Z","lastTransitionTime":"2026-01-22T09:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.018247 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.034338 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.054592 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.074352 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.079343 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.079395 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.079415 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.079439 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.079460 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.098878 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.121474 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.137662 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.153136 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.180722 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.182323 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.182382 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.182402 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.182430 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.182449 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.218990 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.240832 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.265558 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.281531 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.285830 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.285866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.285877 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.285894 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.285906 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.298229 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.316851 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.335885 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:37Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.389528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.389604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.389626 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.389655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.389676 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.492420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.492499 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.492518 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.492550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.492571 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.595691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.595755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.595775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.595801 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.595820 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.701757 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.702033 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.702125 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.702236 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.702331 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.805288 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.805448 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.805487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.805514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.805530 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.908962 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.909006 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.909027 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.909079 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.909096 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:37Z","lastTransitionTime":"2026-01-22T09:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.956686 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 15:25:06.946909767 +0000 UTC Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.996222 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.996307 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.996320 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:37 crc kubenswrapper[4836]: E0122 09:42:37.996408 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:37 crc kubenswrapper[4836]: I0122 09:42:37.996442 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:37 crc kubenswrapper[4836]: E0122 09:42:37.996583 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:37 crc kubenswrapper[4836]: E0122 09:42:37.996681 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:37 crc kubenswrapper[4836]: E0122 09:42:37.996786 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.012460 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.012535 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.012558 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.012585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.012609 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.116138 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.116229 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.116283 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.116311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.116328 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.197766 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:38 crc kubenswrapper[4836]: E0122 09:42:38.197956 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:38 crc kubenswrapper[4836]: E0122 09:42:38.198094 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:42:46.198014276 +0000 UTC m=+49.557782783 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.219928 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.219986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.219999 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.220021 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.220033 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.323334 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.323427 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.323453 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.323490 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.323517 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.426531 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.426616 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.426626 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.426647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.426660 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.529937 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.530008 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.530103 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.530136 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.530159 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.633483 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.633553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.633577 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.633607 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.633629 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.736832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.736895 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.736912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.736935 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.736953 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.840734 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.840787 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.840795 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.840809 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.840818 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.943651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.943704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.943721 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.943745 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.943763 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:38Z","lastTransitionTime":"2026-01-22T09:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:38 crc kubenswrapper[4836]: I0122 09:42:38.957612 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 01:42:54.041106868 +0000 UTC Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.046349 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.046387 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.046398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.046417 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.046430 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.149721 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.149777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.149795 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.149817 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.149836 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.252345 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.252394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.252405 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.252467 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.252484 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.355202 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.355260 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.355276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.355301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.355317 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.458298 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.458364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.458382 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.458407 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.458427 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.561735 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.561790 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.561802 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.561818 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.561829 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.665826 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.665900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.665918 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.665943 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.665963 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.768760 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.768831 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.768849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.768875 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.768896 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.871929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.872026 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.872089 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.872119 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.872136 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.958453 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 20:49:35.74010764 +0000 UTC Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.975116 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.975180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.975197 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.975222 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.975243 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:39Z","lastTransitionTime":"2026-01-22T09:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.996853 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.996949 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:39 crc kubenswrapper[4836]: E0122 09:42:39.997117 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.997150 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:39 crc kubenswrapper[4836]: I0122 09:42:39.997196 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:39 crc kubenswrapper[4836]: E0122 09:42:39.997434 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:39 crc kubenswrapper[4836]: E0122 09:42:39.997550 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.018729 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.089426 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.089512 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.089534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.089570 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.089601 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.193277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.193327 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.193339 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.193359 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.193371 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.296156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.296229 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.296248 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.296274 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.296291 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.400164 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.401424 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.401449 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.401481 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.401503 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.504398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.504461 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.504482 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.504507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.504525 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.607849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.607924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.607958 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.607987 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.608010 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.706525 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.706597 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.706618 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.706644 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.706662 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.724287 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:40Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.729547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.729588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.729600 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.729615 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.729626 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.747037 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:40Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.751553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.751599 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.751615 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.751636 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.751653 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.766268 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:40Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.769558 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.769596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.769610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.769627 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.769638 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.783189 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:40Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.787163 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.787211 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.787227 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.787251 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.787267 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.801695 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:40Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:40 crc kubenswrapper[4836]: E0122 09:42:40.801856 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.803538 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.803585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.803602 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.803623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.803638 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.906441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.906518 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.906534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.906556 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.906572 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:40Z","lastTransitionTime":"2026-01-22T09:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:40 crc kubenswrapper[4836]: I0122 09:42:40.958895 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 05:13:57.511362647 +0000 UTC Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.009912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.009978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.009997 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.010024 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.010068 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.113503 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.113581 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.113621 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.113658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.113682 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.216642 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.216713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.216731 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.216758 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.216776 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.320089 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.320188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.320212 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.320244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.320268 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.423540 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.423605 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.423618 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.423637 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.423649 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.525884 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.525937 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.525956 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.525976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.525991 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.629016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.629096 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.629109 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.629124 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.629135 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.741472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.741547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.741566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.741589 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.741605 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.844767 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.844827 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.844849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.844878 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.844901 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.949244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.949304 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.949324 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.949349 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.949366 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:41Z","lastTransitionTime":"2026-01-22T09:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.959512 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 14:32:17.137391143 +0000 UTC Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.996323 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.996355 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.996427 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.996443 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:41 crc kubenswrapper[4836]: E0122 09:42:41.996516 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:41 crc kubenswrapper[4836]: E0122 09:42:41.996625 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:41 crc kubenswrapper[4836]: E0122 09:42:41.997223 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:41 crc kubenswrapper[4836]: E0122 09:42:41.997303 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:41 crc kubenswrapper[4836]: I0122 09:42:41.997708 4836 scope.go:117] "RemoveContainer" containerID="35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.053259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.053521 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.053534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.053551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.053563 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.157824 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.157866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.157878 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.157893 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.157903 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.260822 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.260890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.260908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.260934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.260950 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.364130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.364187 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.364205 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.364230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.364248 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.467472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.467527 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.467550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.467580 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.467599 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.570478 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.570521 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.570535 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.570551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.570562 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.673463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.673497 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.673505 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.673517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.673525 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.776003 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.776112 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.776136 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.776165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.776183 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.835536 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.879326 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.879395 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.879435 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.879466 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.879490 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.960324 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 11:21:34.236496043 +0000 UTC Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.981864 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.981915 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.981934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.981960 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:42 crc kubenswrapper[4836]: I0122 09:42:42.981980 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:42Z","lastTransitionTime":"2026-01-22T09:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.084855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.084908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.084927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.084950 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.084967 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.188440 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.188507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.188532 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.188563 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.188587 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.291923 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.292008 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.292028 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.292085 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.292107 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.394767 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.394828 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.394842 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.394860 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.394873 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.443833 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/1.log" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.446011 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.446574 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.456170 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.471060 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.481127 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.492999 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.496434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.496479 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.496491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.496509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.496522 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.505150 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.516283 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.548777 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.564921 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.596900 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.598423 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.598464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.598476 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.598491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.598502 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.622314 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.634919 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.647101 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.661187 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.686130 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.698425 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.701350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.701380 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.701388 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.701401 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.701409 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.708694 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:43Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.803246 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.803278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.803286 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.803297 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.803307 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.905907 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.905947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.905955 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.905968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.905977 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:43Z","lastTransitionTime":"2026-01-22T09:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.960970 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 18:01:59.021977585 +0000 UTC Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.996921 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.997111 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.997220 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:43 crc kubenswrapper[4836]: E0122 09:42:43.997216 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:43 crc kubenswrapper[4836]: I0122 09:42:43.997259 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:43 crc kubenswrapper[4836]: E0122 09:42:43.997363 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:43 crc kubenswrapper[4836]: E0122 09:42:43.997456 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:43 crc kubenswrapper[4836]: E0122 09:42:43.997549 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.008635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.008698 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.008709 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.008732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.008745 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.110919 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.110952 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.110961 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.110974 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.110983 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.214165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.214276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.214294 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.214313 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.214331 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.317441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.317477 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.317487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.317504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.317516 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.421132 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.421199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.421212 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.421231 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.421245 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.524183 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.524255 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.524281 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.524312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.524335 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.627533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.627591 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.627608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.627636 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.627655 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.730765 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.730867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.730903 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.730938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.730964 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.834673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.834759 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.834777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.834802 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.834822 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.937446 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.937509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.937525 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.937549 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.937566 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:44Z","lastTransitionTime":"2026-01-22T09:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:44 crc kubenswrapper[4836]: I0122 09:42:44.961879 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 16:23:21.42962018 +0000 UTC Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.039847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.039886 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.039896 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.039914 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.039929 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.143580 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.143632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.143647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.143667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.143681 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.246735 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.246806 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.246831 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.246860 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.246881 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.350205 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.350293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.350318 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.350349 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.350377 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.451928 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.451965 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.451978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.451995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.452006 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.455081 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/2.log" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.455272 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.455882 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/1.log" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.458860 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224" exitCode=1 Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.458900 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.458936 4836 scope.go:117] "RemoveContainer" containerID="35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.459751 4836 scope.go:117] "RemoveContainer" containerID="9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224" Jan 22 09:42:45 crc kubenswrapper[4836]: E0122 09:42:45.459968 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.466651 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.474651 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.488256 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.501489 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.514334 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.524836 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.538531 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.548984 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.555192 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.555230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.555243 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.555258 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.555267 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.562382 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.575929 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.587242 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.599399 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.611945 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.629265 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.646553 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.657724 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.657766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.657779 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.657799 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.657814 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.671716 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.690763 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.707626 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.725117 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.747183 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.760196 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.760253 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.760274 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.760296 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.760311 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.770486 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.796934 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.812487 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.828001 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.840782 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.853388 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.863270 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.863322 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.863337 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.863357 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.863372 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.867541 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.878741 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.892566 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.906851 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.920462 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.935338 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.947521 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.961453 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:45Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.962620 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 23:54:22.46707687 +0000 UTC Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.972095 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.972151 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.972203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.972220 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.972232 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:45Z","lastTransitionTime":"2026-01-22T09:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.996113 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.996166 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.996164 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:45 crc kubenswrapper[4836]: I0122 09:42:45.996302 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:45 crc kubenswrapper[4836]: E0122 09:42:45.996295 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:45 crc kubenswrapper[4836]: E0122 09:42:45.996535 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:45 crc kubenswrapper[4836]: E0122 09:42:45.996668 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:45 crc kubenswrapper[4836]: E0122 09:42:45.996786 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.074984 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.075032 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.075060 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.075076 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.075088 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.177498 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.177553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.177570 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.177589 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.177603 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.280739 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.280812 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.280836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.280900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.280926 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.297107 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:46 crc kubenswrapper[4836]: E0122 09:42:46.297364 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:46 crc kubenswrapper[4836]: E0122 09:42:46.297470 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:02.297440765 +0000 UTC m=+65.657209312 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.385511 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.385587 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.385604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.385628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.385647 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.466349 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/2.log" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.488556 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.488658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.488682 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.488711 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.488766 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.591629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.591698 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.591721 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.591751 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.591769 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.695080 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.695158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.695184 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.695218 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.695236 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.798923 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.799023 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.799074 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.799105 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.799126 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.901748 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.901927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.901960 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.902000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.902024 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:46Z","lastTransitionTime":"2026-01-22T09:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:46 crc kubenswrapper[4836]: I0122 09:42:46.963084 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:49:04.14145365 +0000 UTC Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.004671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.004731 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.004742 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.004759 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.004772 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.014004 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.029258 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.047686 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.061214 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.093259 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.106960 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.107029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.107093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.107126 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.107150 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.107358 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.121825 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.139002 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.156595 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.169743 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.182964 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.203088 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.209440 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.209500 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.209518 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.209543 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.209562 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.222439 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.240343 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.254932 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.269010 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.287101 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:47Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.312059 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.312117 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.312134 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.312156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.312173 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.415629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.416209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.416230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.416255 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.416274 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.518913 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.518973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.518991 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.519014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.519031 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.621730 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.621779 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.621804 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.621825 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.621838 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.725941 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.726010 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.726029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.726091 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.726111 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.829604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.829673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.829691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.829716 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.829733 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.914348 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.914519 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914662 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:19.914610564 +0000 UTC m=+83.274379131 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914666 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.914722 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914768 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:19.914754218 +0000 UTC m=+83.274522775 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.914813 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914851 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914880 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914900 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.914938 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.914971 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:19.914945873 +0000 UTC m=+83.274714420 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915091 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915180 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:19.915151748 +0000 UTC m=+83.274920315 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915218 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915252 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915275 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.915352 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:19.915331063 +0000 UTC m=+83.275099660 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.932537 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.932593 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.932610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.932633 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.932651 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:47Z","lastTransitionTime":"2026-01-22T09:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.963291 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 07:32:29.579334874 +0000 UTC Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.996705 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.996758 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.996707 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:47 crc kubenswrapper[4836]: I0122 09:42:47.996851 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.997033 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.997267 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.997449 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:47 crc kubenswrapper[4836]: E0122 09:42:47.997526 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.035691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.035811 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.035829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.035852 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.035868 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.139203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.139277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.139290 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.139308 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.139321 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.241648 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.241691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.241702 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.241719 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.241730 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.344670 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.344709 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.344718 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.344731 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.344741 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.447152 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.447199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.447211 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.447229 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.447241 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.550109 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.550156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.550179 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.550208 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.550227 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.653553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.653639 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.653663 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.653693 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.653714 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.756790 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.756849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.756866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.756890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.756911 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.860667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.860739 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.860763 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.860792 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.860819 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963424 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 20:55:05.493021807 +0000 UTC Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963761 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963804 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963841 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963859 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:48 crc kubenswrapper[4836]: I0122 09:42:48.963870 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:48Z","lastTransitionTime":"2026-01-22T09:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.066585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.066642 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.066654 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.066674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.066686 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.169968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.170094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.170122 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.170145 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.170163 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.273730 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.273789 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.273806 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.273867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.273887 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.376617 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.376700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.376713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.376730 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.376743 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.482369 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.482433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.482449 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.482472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.482489 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.585731 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.585822 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.585853 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.585904 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.585931 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.689579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.689657 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.689684 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.689714 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.689735 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.792442 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.792543 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.792561 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.792584 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.792600 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.897924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.897983 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.897999 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.898083 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.898104 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:49Z","lastTransitionTime":"2026-01-22T09:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.964245 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 22:12:11.696636341 +0000 UTC Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.996636 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.996681 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.996709 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:49 crc kubenswrapper[4836]: I0122 09:42:49.996783 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:49 crc kubenswrapper[4836]: E0122 09:42:49.996878 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:49 crc kubenswrapper[4836]: E0122 09:42:49.997006 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:49 crc kubenswrapper[4836]: E0122 09:42:49.997123 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:49 crc kubenswrapper[4836]: E0122 09:42:49.997210 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.001406 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.001459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.001481 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.001505 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.001523 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.104511 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.104545 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.104555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.104571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.104583 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.208112 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.208168 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.208190 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.208217 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.208237 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.311219 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.311255 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.311266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.311282 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.311294 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.414409 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.414459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.414470 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.414487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.414497 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.517016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.517089 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.517103 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.517120 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.517134 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.620076 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.620141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.620162 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.620191 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.620211 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.723293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.723352 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.723370 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.723395 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.723413 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.825955 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.826153 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.826186 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.826220 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.826262 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.876599 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.876670 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.876689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.876716 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.876734 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.894830 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:50Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.899565 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.899616 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.899631 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.899655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.899669 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.943233 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:50Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.947595 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.947634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.947645 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.947661 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.947675 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.959252 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:50Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.963298 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.963339 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.963351 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.963369 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.963381 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.964357 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 11:21:47.836431921 +0000 UTC Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.978496 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:50Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.982375 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.982421 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.982437 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.982459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:50 crc kubenswrapper[4836]: I0122 09:42:50.982474 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:50Z","lastTransitionTime":"2026-01-22T09:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.998503 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:50Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:50 crc kubenswrapper[4836]: E0122 09:42:50.998660 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.000506 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.000539 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.000550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.000566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.000577 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.103514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.103553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.103564 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.103579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.103590 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.207238 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.207366 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.207388 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.207411 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.207432 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.310543 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.310604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.310621 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.310643 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.310661 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.414276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.414336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.414350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.414373 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.414391 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.532791 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.532876 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.532894 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.532917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.532936 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.635907 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.636015 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.636035 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.636104 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.636122 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.739815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.739870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.739890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.739912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.739929 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.842840 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.842917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.842938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.842967 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.842992 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.945370 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.945433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.945455 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.945505 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.945525 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:51Z","lastTransitionTime":"2026-01-22T09:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.965216 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:42:18.442213728 +0000 UTC Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.996106 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.996137 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.996192 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:51 crc kubenswrapper[4836]: I0122 09:42:51.996221 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:51 crc kubenswrapper[4836]: E0122 09:42:51.996368 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:51 crc kubenswrapper[4836]: E0122 09:42:51.996474 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:51 crc kubenswrapper[4836]: E0122 09:42:51.996605 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:51 crc kubenswrapper[4836]: E0122 09:42:51.996690 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.048192 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.048291 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.048310 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.048336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.048354 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.151401 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.151477 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.151500 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.151527 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.151550 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.254645 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.254700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.254711 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.254724 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.254733 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.357388 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.357468 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.357491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.357522 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.357544 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.460598 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.460661 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.460683 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.460706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.460723 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.563247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.563302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.563311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.563328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.563338 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.666414 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.666484 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.666507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.666539 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.666562 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.769098 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.769160 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.769177 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.769202 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.769221 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.871809 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.871864 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.871881 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.871904 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.871925 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.965654 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 09:28:25.672659323 +0000 UTC Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.974910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.974961 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.974972 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.974993 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:52 crc kubenswrapper[4836]: I0122 09:42:52.975008 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:52Z","lastTransitionTime":"2026-01-22T09:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.077405 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.077458 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.077476 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.077503 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.077702 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.180803 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.180849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.180859 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.180880 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.180892 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.284196 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.284254 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.284272 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.284297 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.284316 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.387944 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.388000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.388017 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.388071 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.388090 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.490937 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.490988 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.490996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.491071 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.491083 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.594923 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.594997 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.595020 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.595080 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.595100 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.698681 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.698776 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.698794 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.698823 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.698840 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.802309 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.802422 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.802439 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.802465 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.802484 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.905584 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.905761 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.905787 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.905815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.905838 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:53Z","lastTransitionTime":"2026-01-22T09:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.966788 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 08:28:55.532246989 +0000 UTC Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.996084 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.996117 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.996104 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:53 crc kubenswrapper[4836]: I0122 09:42:53.996091 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:53 crc kubenswrapper[4836]: E0122 09:42:53.996211 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:53 crc kubenswrapper[4836]: E0122 09:42:53.996311 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:53 crc kubenswrapper[4836]: E0122 09:42:53.996543 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:53 crc kubenswrapper[4836]: E0122 09:42:53.996690 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.009067 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.009131 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.009149 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.009176 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.009195 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.112732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.112780 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.112815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.112834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.112845 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.216276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.216341 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.216359 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.216384 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.216404 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.319116 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.319400 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.319519 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.319635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.319753 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.422358 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.422418 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.422434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.422463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.422488 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.525279 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.525329 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.525347 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.525371 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.525390 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.628117 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.628198 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.628222 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.628252 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.628273 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.730781 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.730848 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.730882 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.730941 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.730969 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.833613 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.833678 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.833694 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.833717 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.833733 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.936706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.936766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.936784 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.936813 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.936832 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:54Z","lastTransitionTime":"2026-01-22T09:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:54 crc kubenswrapper[4836]: I0122 09:42:54.967102 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:42:02.636012426 +0000 UTC Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.040444 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.040502 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.040519 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.040540 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.040552 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.142958 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.143079 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.143109 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.143140 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.143163 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.246275 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.246364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.246383 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.246406 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.246422 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.349003 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.349082 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.349118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.349136 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.349146 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.451725 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.451766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.451776 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.451793 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.451804 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.553778 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.553816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.553824 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.553837 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.553845 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.657455 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.657546 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.657565 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.657590 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.657606 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.760618 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.760665 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.760674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.760689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.760697 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.868082 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.868128 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.868139 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.868155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.868166 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.967895 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 22:58:13.818527243 +0000 UTC Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.971230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.971287 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.971305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.971328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.971345 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:55Z","lastTransitionTime":"2026-01-22T09:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.996774 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.996842 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.996925 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:55 crc kubenswrapper[4836]: E0122 09:42:55.996923 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:55 crc kubenswrapper[4836]: I0122 09:42:55.996987 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:55 crc kubenswrapper[4836]: E0122 09:42:55.997149 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:55 crc kubenswrapper[4836]: E0122 09:42:55.997228 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:55 crc kubenswrapper[4836]: E0122 09:42:55.997326 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.074299 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.074348 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.074359 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.074376 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.074387 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.177272 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.177336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.177353 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.177379 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.177397 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.280325 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.280381 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.280398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.280417 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.280432 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.383312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.383364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.383376 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.383389 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.383400 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.486640 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.486696 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.486714 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.486737 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.486757 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.589157 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.589218 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.589241 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.589269 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.589292 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.692407 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.692448 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.692457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.692475 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.692485 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.796204 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.796262 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.796278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.796305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.796323 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.899141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.899238 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.899264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.899299 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.899324 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:56Z","lastTransitionTime":"2026-01-22T09:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.968854 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 22:46:14.401858265 +0000 UTC Jan 22 09:42:56 crc kubenswrapper[4836]: I0122 09:42:56.996894 4836 scope.go:117] "RemoveContainer" containerID="9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224" Jan 22 09:42:56 crc kubenswrapper[4836]: E0122 09:42:56.997139 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.002315 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.002373 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.002391 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.002424 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.002443 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.024290 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35464f7aa58edfa15905436b89a08cf8d71f17c82fddb94a1208517c15fdf22a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:27Z\\\",\\\"message\\\":\\\" Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0122 09:42:27.884129 6226 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: could not add Event Handler for podInformer during egressqosController initialization, handler {0x21cf880 0x21cf560 0x21cf500} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:27Z is after 2025-08-24T17:21:41Z]\\\\nI0122 09:42:27.884641 6226 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-operator cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.042780 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.057126 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.078880 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.102662 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.105184 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.105248 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.105272 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.105302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.105323 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.123706 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.141777 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.162202 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.182265 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.199637 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.209094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.209141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.209155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.209175 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.209193 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.218191 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.233951 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.248448 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.263973 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.283983 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.302366 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.311917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.311962 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.311974 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.311992 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.312008 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.327494 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.352376 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.373567 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.392339 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.409540 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.414608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.415024 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.415056 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.415081 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.415096 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.422187 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.436535 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.449668 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.467648 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.489198 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.513485 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.517156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.517199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.517211 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.517264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.517283 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.538554 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.553448 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.569826 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.587857 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.599670 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.611908 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.619520 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.619571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.619584 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.619603 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.619614 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.623387 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:42:57Z is after 2025-08-24T17:21:41Z" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.722757 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.722807 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.722816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.722832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.722844 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.824772 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.824846 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.824863 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.824882 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.824894 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.927506 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.927543 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.927554 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.927568 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.927577 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:57Z","lastTransitionTime":"2026-01-22T09:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.969901 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 00:23:53.076019389 +0000 UTC Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.996242 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:57 crc kubenswrapper[4836]: E0122 09:42:57.996383 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.996248 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.996446 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:57 crc kubenswrapper[4836]: E0122 09:42:57.996650 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:57 crc kubenswrapper[4836]: E0122 09:42:57.996466 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:57 crc kubenswrapper[4836]: I0122 09:42:57.996769 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:57 crc kubenswrapper[4836]: E0122 09:42:57.996825 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.030632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.030728 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.030744 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.030768 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.030782 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.140400 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.140450 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.140464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.140482 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.140496 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.242756 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.242829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.242847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.242953 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.243016 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.345350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.345441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.345483 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.345503 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.345517 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.448511 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.448595 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.448615 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.448640 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.448662 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.551634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.551706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.551717 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.551737 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.551749 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.655012 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.655083 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.655093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.655113 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.655124 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.759076 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.759143 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.759155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.759189 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.759204 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.861953 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.862011 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.862024 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.862062 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.862078 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.964846 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.964887 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.964898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.964916 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.964927 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:58Z","lastTransitionTime":"2026-01-22T09:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:58 crc kubenswrapper[4836]: I0122 09:42:58.971116 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 17:07:52.952675564 +0000 UTC Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.068411 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.068499 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.068516 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.068554 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.068577 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.177623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.177691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.177710 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.177789 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.177814 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.281753 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.281825 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.281840 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.281868 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.281883 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.384973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.385373 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.385541 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.385683 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.385824 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.488343 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.488391 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.488403 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.488420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.488431 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.591232 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.591314 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.591338 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.591371 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.591394 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.694338 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.694373 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.694381 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.694394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.694403 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.797507 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.797574 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.797596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.797620 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.797637 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.899799 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.899853 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.899867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.899884 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.899896 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:42:59Z","lastTransitionTime":"2026-01-22T09:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.972153 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 13:52:56.277370508 +0000 UTC Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.996637 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.996799 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:42:59 crc kubenswrapper[4836]: E0122 09:42:59.996985 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.997074 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:42:59 crc kubenswrapper[4836]: E0122 09:42:59.997154 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:42:59 crc kubenswrapper[4836]: E0122 09:42:59.997376 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:42:59 crc kubenswrapper[4836]: I0122 09:42:59.997507 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:42:59 crc kubenswrapper[4836]: E0122 09:42:59.997640 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.003167 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.003241 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.003266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.003301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.003326 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.107252 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.107332 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.107350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.107375 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.107399 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.210676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.210761 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.211153 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.211461 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.211790 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.314430 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.314476 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.314487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.314504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.314517 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.416344 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.416382 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.416391 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.416404 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.416413 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.518978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.519024 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.519057 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.519077 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.519089 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.621270 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.621338 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.621374 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.621403 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.621424 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.724768 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.724806 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.724815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.724829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.724839 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.827976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.828080 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.828095 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.828114 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.828126 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.930169 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.930213 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.930225 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.930242 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.930255 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:00Z","lastTransitionTime":"2026-01-22T09:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:00 crc kubenswrapper[4836]: I0122 09:43:00.972711 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 21:20:29.34515017 +0000 UTC Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.033032 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.033090 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.033100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.033115 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.033126 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.083706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.083741 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.083749 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.083762 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.083772 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.096909 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:01Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.102876 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.102910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.102922 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.102938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.102950 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.115313 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:01Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.119144 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.119173 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.119184 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.119201 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.119214 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.131487 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:01Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.134529 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.134572 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.134583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.134601 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.134616 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.147550 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:01Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.151676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.151708 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.151720 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.151736 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.151749 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.164902 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:01Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.165069 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.166659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.166679 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.166691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.166706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.166718 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.269276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.269336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.269348 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.269364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.269376 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.371301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.371333 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.371341 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.371354 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.371363 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.472947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.473299 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.473448 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.473729 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.473938 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.577060 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.577108 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.577119 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.577135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.577148 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.680489 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.680517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.680550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.680564 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.680572 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.783484 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.783536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.783548 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.783571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.783585 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.886374 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.886440 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.886465 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.886495 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.886521 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.973611 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 01:11:13.822057502 +0000 UTC Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.989159 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.989199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.989209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.989230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.989241 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:01Z","lastTransitionTime":"2026-01-22T09:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.996732 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.996744 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.996816 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.996853 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.996905 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:01 crc kubenswrapper[4836]: I0122 09:43:01.996926 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.996949 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:01 crc kubenswrapper[4836]: E0122 09:43:01.997087 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.091912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.091947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.091956 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.092018 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.092030 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.194228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.194281 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.194303 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.194322 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.194335 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.297231 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.297280 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.297292 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.297311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.297325 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.363278 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:02 crc kubenswrapper[4836]: E0122 09:43:02.363390 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:43:02 crc kubenswrapper[4836]: E0122 09:43:02.363440 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:34.363425782 +0000 UTC m=+97.723194299 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.399570 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.399621 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.399629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.399643 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.399652 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.501897 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.501938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.501946 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.501964 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.501973 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.604335 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.604375 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.604385 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.604399 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.604410 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.706849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.706898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.706913 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.706962 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.706977 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.810170 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.810237 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.810250 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.810287 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.810299 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.913396 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.913674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.913682 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.913699 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.913708 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:02Z","lastTransitionTime":"2026-01-22T09:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:02 crc kubenswrapper[4836]: I0122 09:43:02.974574 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 00:26:27.743883184 +0000 UTC Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.015583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.015630 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.015641 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.015659 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.015671 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.118430 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.118521 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.118540 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.118577 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.118594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.221442 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.221482 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.221493 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.221509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.221520 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.323927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.323976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.323988 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.324006 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.324018 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.426890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.427142 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.427204 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.427301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.427361 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.529992 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.530035 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.530061 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.530075 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.530084 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.532376 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/0.log" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.532420 4836 generic.go:334] "Generic (PLEG): container finished" podID="7fcb3e24-f305-4f39-af85-2dc05a0af79f" containerID="07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784" exitCode=1 Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.532462 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerDied","Data":"07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.532919 4836 scope.go:117] "RemoveContainer" containerID="07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.548923 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.563002 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.576062 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.591721 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.603262 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.613619 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.627915 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.632632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.632666 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.632676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.632689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.632699 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.639576 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.651142 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.662192 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.672796 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.682451 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.698466 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.711224 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.725142 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.735158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.735206 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.735219 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.735238 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.735250 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.746379 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.758233 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:03Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.838313 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.838368 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.838405 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.838434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.838454 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.941250 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.941293 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.941302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.941319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.941330 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:03Z","lastTransitionTime":"2026-01-22T09:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.975185 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 12:06:44.567831698 +0000 UTC Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.996757 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.996838 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:03 crc kubenswrapper[4836]: E0122 09:43:03.996903 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:03 crc kubenswrapper[4836]: E0122 09:43:03.997210 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.997273 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:03 crc kubenswrapper[4836]: I0122 09:43:03.997282 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:03 crc kubenswrapper[4836]: E0122 09:43:03.997353 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:03 crc kubenswrapper[4836]: E0122 09:43:03.997428 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.047952 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.047986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.047998 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.048014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.048025 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.150873 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.150907 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.150917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.150932 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.150942 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.253423 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.253459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.253470 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.253485 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.253494 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.356103 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.356137 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.356149 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.356164 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.356174 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.459002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.459063 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.459074 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.459089 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.459101 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.561362 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.561412 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.561424 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.561441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.561454 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.663924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.663985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.663995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.664015 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.664028 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.766646 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.766685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.766697 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.766713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.766725 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.869521 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.869562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.869573 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.869588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.869600 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.972145 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.972179 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.972190 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.972203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.972212 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:04Z","lastTransitionTime":"2026-01-22T09:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:04 crc kubenswrapper[4836]: I0122 09:43:04.975684 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 21:28:47.332550454 +0000 UTC Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.074596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.074631 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.074640 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.074655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.074666 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.176953 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.176989 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.176998 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.177012 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.177023 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.278922 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.278966 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.278978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.278995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.279011 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.382207 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.382263 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.382279 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.382297 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.382313 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.484594 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.484636 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.484647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.484664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.484674 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.540455 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/0.log" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.540528 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerStarted","Data":"023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.554357 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.567843 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.580401 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.587533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.587576 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.587588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.587610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.587625 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.593826 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.605576 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.614822 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.627953 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.640790 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.654616 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.675296 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.688944 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.689632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.689661 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.689671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.689688 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.689700 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.704865 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.718485 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.736330 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.753925 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.770298 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.784432 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:05Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.792780 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.792832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.792843 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.792862 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.792876 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.895180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.895229 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.895245 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.895268 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.895284 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.976724 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 18:23:33.219882596 +0000 UTC Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.996163 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.996197 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.996210 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:05 crc kubenswrapper[4836]: E0122 09:43:05.996346 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.996364 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:05 crc kubenswrapper[4836]: E0122 09:43:05.996476 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:05 crc kubenswrapper[4836]: E0122 09:43:05.996699 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:05 crc kubenswrapper[4836]: E0122 09:43:05.996779 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.998203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.998236 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.998248 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.998266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:05 crc kubenswrapper[4836]: I0122 09:43:05.998278 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:05Z","lastTransitionTime":"2026-01-22T09:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.100674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.100723 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.100736 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.100753 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.100765 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.203031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.203104 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.203120 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.203142 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.203156 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.305964 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.306006 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.306017 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.306033 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.306063 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.408866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.408910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.408920 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.408935 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.408948 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.510879 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.510924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.510934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.510951 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.510965 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.612862 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.612904 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.612916 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.612933 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.612953 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.715082 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.715109 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.715118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.715133 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.715144 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.818196 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.818264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.818277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.818294 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.818306 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.920939 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.920985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.920996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.921013 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.921022 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:06Z","lastTransitionTime":"2026-01-22T09:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:06 crc kubenswrapper[4836]: I0122 09:43:06.977491 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 01:19:18.063875796 +0000 UTC Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.005830 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.011277 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023070 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023597 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.023631 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.033684 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.042971 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.056333 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.067512 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.079994 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.089940 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.102385 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.116011 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.129612 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.129762 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.129815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.129869 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.129900 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.140074 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.155029 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.168753 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.196880 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.207906 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.219506 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.230432 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:07Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.234079 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.234135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.234156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.234187 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.234210 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.336094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.336134 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.336145 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.336160 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.336171 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.438524 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.438570 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.438582 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.438599 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.438612 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.540652 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.540678 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.540687 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.540700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.540709 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.643266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.643309 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.643320 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.643335 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.643347 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.745668 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.745706 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.745716 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.745732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.745742 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.848349 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.848389 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.848400 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.848418 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.848439 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.950951 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.950991 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.951002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.951017 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.951025 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:07Z","lastTransitionTime":"2026-01-22T09:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.978262 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 12:50:25.920879325 +0000 UTC Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.996870 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.996936 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.996933 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.996901 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:07 crc kubenswrapper[4836]: E0122 09:43:07.997085 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:07 crc kubenswrapper[4836]: E0122 09:43:07.997191 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:07 crc kubenswrapper[4836]: E0122 09:43:07.997269 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:07 crc kubenswrapper[4836]: E0122 09:43:07.997665 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:07 crc kubenswrapper[4836]: I0122 09:43:07.997957 4836 scope.go:117] "RemoveContainer" containerID="9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.053241 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.053504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.053517 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.053534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.053546 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.155501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.155523 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.155533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.155551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.155564 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.258199 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.258235 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.258244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.258259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.258269 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.361268 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.361326 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.361338 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.361358 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.361372 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.467766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.468118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.468132 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.468157 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.468170 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.550994 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/2.log" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.553659 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.554061 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.566070 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.570092 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.570126 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.570135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.570149 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.570158 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.579673 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.592958 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.606403 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.621201 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.632535 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.645069 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.657993 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.671960 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.672000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.672014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.672029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.672062 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.675526 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.689359 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.709441 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.720901 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.732959 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.745007 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.756367 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506fe75f-8c00-4768-a50b-80dd7afde450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.767124 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.774240 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.774269 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.774278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.774291 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.774298 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.781110 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.791454 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:08Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.876742 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.876778 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.876788 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.876802 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.876812 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.978529 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 01:18:44.030363744 +0000 UTC Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.979565 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.979623 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.979643 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.979670 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:08 crc kubenswrapper[4836]: I0122 09:43:08.979691 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:08Z","lastTransitionTime":"2026-01-22T09:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.082098 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.082147 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.082160 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.082180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.082191 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.184700 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.184730 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.184738 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.184751 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.184760 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.287657 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.287717 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.287735 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.287758 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.287777 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.390832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.390874 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.390886 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.390905 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.390918 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.492947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.492984 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.492992 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.493007 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.493017 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.595820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.595872 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.595884 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.595902 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.595919 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.697836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.697886 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.697897 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.697915 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.697930 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.800684 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.800745 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.800762 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.800787 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.800804 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.903233 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.903319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.903358 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.903380 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.903392 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:09Z","lastTransitionTime":"2026-01-22T09:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.980085 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 22:06:24.821597339 +0000 UTC Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.996469 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.996537 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.996535 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:09 crc kubenswrapper[4836]: E0122 09:43:09.996725 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:09 crc kubenswrapper[4836]: E0122 09:43:09.996865 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:09 crc kubenswrapper[4836]: I0122 09:43:09.997020 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:09 crc kubenswrapper[4836]: E0122 09:43:09.997019 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:09 crc kubenswrapper[4836]: E0122 09:43:09.997229 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.006410 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.006441 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.006450 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.006464 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.006473 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.109266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.109312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.109321 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.109336 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.109346 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.212093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.212135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.212147 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.212165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.212177 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.314498 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.314547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.314562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.314580 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.314594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.416986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.417072 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.417099 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.417132 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.417156 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.520315 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.520646 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.520819 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.521645 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.521696 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.561349 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/3.log" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.562236 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/2.log" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.565438 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" exitCode=1 Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.565486 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.565538 4836 scope.go:117] "RemoveContainer" containerID="9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.566245 4836 scope.go:117] "RemoveContainer" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" Jan 22 09:43:10 crc kubenswrapper[4836]: E0122 09:43:10.566389 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.581356 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.593603 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.603800 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.617194 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506fe75f-8c00-4768-a50b-80dd7afde450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.623889 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.623971 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.623996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.624026 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.624086 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.633260 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.649631 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.669120 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.687192 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.700094 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.714944 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.726230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.726266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.726277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.726292 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.726304 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.728482 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.750523 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.767769 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.798596 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:09Z\\\",\\\"message\\\":\\\"eSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},ServicePort{Name:health,Protocol:TCP,Port:8798,TargetPort:{0 8798 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-daemon,},ClusterIP:10.217.4.43,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.43],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 09:43:08.875072 6834 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-daemon for network=default are: map[]\\\\nI0122 09:43:08.875072 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-7wkg5\\\\nI0122 09:43:08.875096 6834 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-7wkg5 in node crc\\\\nF0122 09:43:08.875133 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.812722 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.828793 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.829033 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.829088 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.829100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.829118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.829131 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.844489 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.863085 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:10Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.931832 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.931894 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.931911 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.931938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.931957 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:10Z","lastTransitionTime":"2026-01-22T09:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:10 crc kubenswrapper[4836]: I0122 09:43:10.980568 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 18:52:15.240496551 +0000 UTC Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.036996 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.037411 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.037551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.037692 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.037828 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.141331 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.141408 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.141428 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.141457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.141477 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.236875 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.236983 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.237004 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.237029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.237098 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.256943 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:11Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.262359 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.262596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.262753 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.262903 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.263089 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.283700 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:11Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.288772 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.288839 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.288858 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.288882 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.288899 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.309605 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:11Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.315634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.315918 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.316134 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.316316 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.316457 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.340750 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:11Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.345929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.346272 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.346465 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.346627 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.346770 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.369133 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:11Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.369861 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.372428 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.372641 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.372792 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.373002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.373296 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.476667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.476713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.476725 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.476742 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.476775 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.571842 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/3.log" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.579394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.579440 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.579457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.579479 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.579496 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.682502 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.682551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.682572 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.682593 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.682606 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.785328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.785400 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.785427 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.785457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.785478 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.888078 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.888125 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.888138 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.888156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.888168 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.981775 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 15:02:01.309292039 +0000 UTC Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.990460 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.990509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.990520 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.990536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.990548 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.996311 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.996380 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.996409 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.996418 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:11 crc kubenswrapper[4836]: I0122 09:43:11.996463 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.996577 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.996643 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:11 crc kubenswrapper[4836]: E0122 09:43:11.996830 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.093822 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.093874 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.093890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.093913 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.093930 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.197501 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.197551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.197562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.197580 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.197594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.300389 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.300442 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.300456 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.300476 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.300489 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.402729 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.402777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.402790 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.402807 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.402819 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.505014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.505124 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.505161 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.505192 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.505213 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.608392 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.608452 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.608466 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.608484 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.608496 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.710612 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.710653 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.710664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.710681 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.710691 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.813592 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.813667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.813689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.813719 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.813742 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.917013 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.917128 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.917152 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.917183 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.917263 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:12 crc kubenswrapper[4836]: I0122 09:43:12.982654 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 09:04:11.770238716 +0000 UTC Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.019335 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.019369 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.019378 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.019407 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.019417 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.121634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.121686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.121697 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.121713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.121726 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.224540 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.224614 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.224625 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.224644 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.224657 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.327349 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.327403 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.327413 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.327433 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.327450 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.430150 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.430192 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.430203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.430221 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.430233 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.532676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.532738 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.532755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.532772 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.532788 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.635383 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.635438 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.635457 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.635483 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.635499 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.737966 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.738036 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.738096 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.738135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.738162 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.841773 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.841845 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.841863 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.841885 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.841904 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.945487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.945545 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.945559 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.945579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.945592 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.982975 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 15:27:26.856896671 +0000 UTC Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.996712 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.996736 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.996794 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:13 crc kubenswrapper[4836]: E0122 09:43:13.996824 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:13 crc kubenswrapper[4836]: I0122 09:43:13.996852 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:13 crc kubenswrapper[4836]: E0122 09:43:13.996932 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:13 crc kubenswrapper[4836]: E0122 09:43:13.997109 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:13 crc kubenswrapper[4836]: E0122 09:43:13.997211 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.048867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.048932 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.048955 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.048984 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.049007 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.151755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.151834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.151858 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.151892 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.151917 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.255012 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.255074 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.255108 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.255123 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.255132 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.358205 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.358278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.358301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.358331 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.358352 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.461305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.461365 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.461382 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.461406 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.461424 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.564728 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.564812 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.564844 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.564873 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.564893 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.667434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.667499 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.667522 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.667550 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.667572 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.771330 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.771405 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.771438 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.771468 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.771487 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.874235 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.874286 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.874300 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.874321 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.874335 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.976949 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.976990 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.977000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.977016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.977028 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:14 crc kubenswrapper[4836]: I0122 09:43:14.983911 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 00:30:11.191980463 +0000 UTC Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.080314 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.080376 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.080387 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.080407 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.080420 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.183069 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.183135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.183147 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.183167 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.183178 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.287100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.287208 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.287252 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.287286 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.287330 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.390385 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.390447 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.390460 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.390474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.390482 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.493493 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.493536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.493545 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.493558 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.493566 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.596234 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.596328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.596362 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.596398 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.596424 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.699146 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.699209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.699232 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.699262 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.699282 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.802362 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.802431 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.802449 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.802474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.802493 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.906505 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.906553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.906565 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.906583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.906594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.984787 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 14:09:56.217550058 +0000 UTC Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.996164 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.996220 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.996237 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:15 crc kubenswrapper[4836]: I0122 09:43:15.996299 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:15 crc kubenswrapper[4836]: E0122 09:43:15.996475 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:15 crc kubenswrapper[4836]: E0122 09:43:15.996880 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:15 crc kubenswrapper[4836]: E0122 09:43:15.997209 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:15 crc kubenswrapper[4836]: E0122 09:43:15.997446 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.009156 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.009186 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.009198 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.009218 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.009233 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.113989 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.114075 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.114094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.114130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.114147 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.217506 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.217551 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.217563 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.217581 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.217594 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.319934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.319968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.319977 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.319990 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.319998 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.422650 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.422691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.422704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.422748 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.422762 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.526494 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.526545 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.526555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.526573 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.526586 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.630607 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.630686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.630709 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.630754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.630777 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.734766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.734850 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.734876 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.734906 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.734927 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.838531 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.838601 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.838625 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.838655 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.838680 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.940775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.940829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.940847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.940872 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.940891 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:16 crc kubenswrapper[4836]: I0122 09:43:16.985830 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 19:09:13.605398887 +0000 UTC Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.011775 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.036671 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.043542 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.043602 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.043619 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.043643 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.043661 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.057020 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.094994 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c4150f70fbbb4f7c1e55f0f3e6859df5eac10b72011f01f9ba532e71a1c0224\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:42:44Z\\\",\\\"message\\\":\\\"l:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:42:44.050940 6440 services_controller.go:434] Service openshift-machine-config-operator/machine-config-operator retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{machine-config-operator openshift-machine-config-operator 8bc1afc2-8724-4135-84df-aee09f23af4c 4514 0 2025-02-23 05:12:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:machine-config-operator] map[include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:mco-proxy-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0076fe9db \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-operator,},ClusterIP:10.217\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:09Z\\\",\\\"message\\\":\\\"eSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},ServicePort{Name:health,Protocol:TCP,Port:8798,TargetPort:{0 8798 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-daemon,},ClusterIP:10.217.4.43,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.43],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 09:43:08.875072 6834 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-daemon for network=default are: map[]\\\\nI0122 09:43:08.875072 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-7wkg5\\\\nI0122 09:43:08.875096 6834 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-7wkg5 in node crc\\\\nF0122 09:43:08.875133 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.116276 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.137994 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.146801 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.146966 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.146991 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.147018 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.147036 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.163293 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.194011 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.218862 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.240293 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.250628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.250684 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.250702 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.250728 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.250748 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.257567 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.274828 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506fe75f-8c00-4768-a50b-80dd7afde450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.298623 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.315661 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.335210 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.353479 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.353548 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.353571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.353596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.353616 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.355279 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.372900 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.389147 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.457141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.457209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.457230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.457257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.457275 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.560124 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.560202 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.560225 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.560247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.560264 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.663780 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.663830 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.663848 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.663873 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.663890 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.767114 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.767169 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.767193 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.767219 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.767238 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.870394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.872108 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.872189 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.872225 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.872263 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.975790 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.975828 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.975836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.975850 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.975860 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.987295 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 18:00:05.657083704 +0000 UTC Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.996666 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.996758 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.996788 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:17 crc kubenswrapper[4836]: E0122 09:43:17.996840 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:17 crc kubenswrapper[4836]: I0122 09:43:17.996865 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:17 crc kubenswrapper[4836]: E0122 09:43:17.997066 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:17 crc kubenswrapper[4836]: E0122 09:43:17.997243 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:17 crc kubenswrapper[4836]: E0122 09:43:17.997364 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.078468 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.078618 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.078645 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.078673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.078689 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.181754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.181816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.181829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.181847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.181859 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.284030 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.284093 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.284104 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.284121 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.284133 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.387323 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.387403 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.387422 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.387444 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.387462 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.490754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.490843 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.490870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.490898 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.490916 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.594602 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.594652 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.594664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.594683 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.594696 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.697073 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.697132 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.697149 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.697171 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.697190 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.799649 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.799705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.799722 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.799743 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.799758 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.901722 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.901777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.901794 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.901815 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.901834 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:18 crc kubenswrapper[4836]: I0122 09:43:18.987485 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 17:16:00.517381058 +0000 UTC Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.004328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.004460 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.004473 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.004491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.004507 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.106834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.106876 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.106888 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.106904 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.106915 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.209836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.209901 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.209912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.209948 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.209962 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.313918 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.314000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.314018 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.314086 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.314107 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.416901 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.416970 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.416988 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.417016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.417034 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.518763 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.518805 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.518816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.518834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.518845 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.622704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.622887 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.622914 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.622944 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.622968 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.726654 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.726733 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.726752 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.726781 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.726799 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.830339 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.830553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.830563 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.830576 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.830585 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.932874 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.932911 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.932919 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.932933 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.932942 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.988316 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:59:26.785079242 +0000 UTC Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.994925 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.995023 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.995091 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.995137 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995169 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.995135677 +0000 UTC m=+147.354904234 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995213 4836 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995263 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.99524883 +0000 UTC m=+147.355017347 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.995258 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995314 4836 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995384 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995403 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995420 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995430 4836 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995444 4836 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995448 4836 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995426 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.995399594 +0000 UTC m=+147.355168141 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995517 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.995502406 +0000 UTC m=+147.355270953 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.995552 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.995541667 +0000 UTC m=+147.355310224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.996021 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.996068 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.996072 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:19 crc kubenswrapper[4836]: I0122 09:43:19.996177 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.996192 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.996298 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.996383 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:19 crc kubenswrapper[4836]: E0122 09:43:19.996451 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.036678 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.036732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.036750 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.036775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.036794 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.139484 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.139535 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.139553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.139575 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.139592 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.241969 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.242004 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.242015 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.242031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.242059 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.344537 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.344578 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.344591 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.344608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.344619 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.446719 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.446820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.446839 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.446861 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.446878 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.550756 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.550838 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.550855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.550884 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.550904 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.655849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.655984 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.656119 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.656158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.656238 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.759635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.759666 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.759673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.759686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.759695 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.862490 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.862530 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.862539 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.862552 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.862563 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.966329 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.966488 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.966505 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.966528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.966544 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:20 crc kubenswrapper[4836]: I0122 09:43:20.988613 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 20:31:50.844585059 +0000 UTC Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.068947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.069077 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.069096 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.069116 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.069131 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.171624 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.171667 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.171681 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.171701 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.171718 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.274319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.274367 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.274381 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.274397 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.274409 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.379190 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.379232 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.379243 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.379261 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.379273 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.481953 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.481999 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.482011 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.482026 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.482053 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.584172 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.584266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.584287 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.584311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.584329 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.654002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.654141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.654155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.654181 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.654196 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.671191 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.676555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.676621 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.676638 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.676661 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.676678 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.694531 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.703315 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.703375 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.703385 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.703401 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.703412 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.722109 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.726595 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.726657 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.726671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.726710 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.726723 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.741670 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.747495 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.747745 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.747899 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.748077 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.748238 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.764956 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.765526 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.767455 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.767487 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.767497 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.767513 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.767524 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.870491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.870526 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.870536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.870552 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.870563 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.974631 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.974683 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.974696 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.974715 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.974731 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.989481 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 04:25:40.001976844 +0000 UTC Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.996945 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.997128 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.997341 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.997390 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:21 crc kubenswrapper[4836]: I0122 09:43:21.997433 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.997676 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.997799 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:21 crc kubenswrapper[4836]: E0122 09:43:21.997938 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.079030 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.079105 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.079115 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.079131 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.079142 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.182562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.182611 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.182629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.182651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.182668 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.286555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.286617 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.286634 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.286660 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.286680 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.390480 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.390528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.390545 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.390567 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.390585 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.493214 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.493278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.493296 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.493319 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.493336 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.596480 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.596529 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.596547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.596569 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.596588 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.699942 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.700019 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.700034 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.700089 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.700104 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.802741 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.802809 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.802833 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.802865 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.802889 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.904903 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.904961 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.904978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.905002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.905019 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:22 crc kubenswrapper[4836]: I0122 09:43:22.990190 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 11:52:49.452911776 +0000 UTC Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.006845 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.006873 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.006880 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.006891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.006899 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.109626 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.109927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.110148 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.110311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.110457 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.213817 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.214419 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.214463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.214485 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.214504 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.321523 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.321586 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.321602 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.321629 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.321646 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.424978 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.425099 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.425118 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.425140 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.425158 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.527668 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.527896 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.527905 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.527917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.527927 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.629777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.629811 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.629820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.629836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.629848 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.732778 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.732823 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.732837 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.732856 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.732871 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.834943 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.834991 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.835002 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.835017 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.835028 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.937877 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.937930 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.937940 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.937957 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.937967 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.990514 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 23:38:26.000287115 +0000 UTC Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.996891 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.996923 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.996923 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:23 crc kubenswrapper[4836]: I0122 09:43:23.996975 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:23 crc kubenswrapper[4836]: E0122 09:43:23.997084 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:23 crc kubenswrapper[4836]: E0122 09:43:23.997182 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:23 crc kubenswrapper[4836]: E0122 09:43:23.997264 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:23 crc kubenswrapper[4836]: E0122 09:43:23.997317 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.041463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.042154 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.042193 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.042223 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.042241 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.144205 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.144253 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.144264 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.144282 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.144293 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.246850 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.246910 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.246929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.246952 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.246971 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.349203 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.349273 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.349289 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.349312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.349330 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.450866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.450900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.450908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.450920 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.450931 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.553644 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.553705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.553720 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.553740 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.553754 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.656892 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.656957 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.656975 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.656999 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.657017 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.759301 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.759377 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.759394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.759417 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.759436 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.861985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.862095 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.862125 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.862206 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.862235 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.965067 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.965115 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.965130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.965148 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.965159 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.991282 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 19:12:54.86518276 +0000 UTC Jan 22 09:43:24 crc kubenswrapper[4836]: I0122 09:43:24.997422 4836 scope.go:117] "RemoveContainer" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" Jan 22 09:43:24 crc kubenswrapper[4836]: E0122 09:43:24.997625 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.010964 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.026009 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.040407 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.050617 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.066648 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.067434 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.067476 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.067491 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.067533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.067551 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.081725 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.093650 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.108723 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:09Z\\\",\\\"message\\\":\\\"eSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},ServicePort{Name:health,Protocol:TCP,Port:8798,TargetPort:{0 8798 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-daemon,},ClusterIP:10.217.4.43,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.43],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 09:43:08.875072 6834 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-daemon for network=default are: map[]\\\\nI0122 09:43:08.875072 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-7wkg5\\\\nI0122 09:43:08.875096 6834 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-7wkg5 in node crc\\\\nF0122 09:43:08.875133 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.118579 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.127566 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.138803 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.150978 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.161646 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.170182 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.170230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.170244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.170262 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.170275 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.174370 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.186797 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.198960 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506fe75f-8c00-4768-a50b-80dd7afde450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.211173 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.223056 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.272394 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.272451 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.272461 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.272480 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.272493 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.375527 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.375571 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.375581 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.375603 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.375614 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.478155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.478208 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.478222 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.478243 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.478259 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.580610 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.580647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.580658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.580673 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.580683 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.683772 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.683819 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.683828 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.683841 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.683852 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.786220 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.786276 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.786289 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.786307 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.786321 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.895186 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.895239 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.895253 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.895278 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.895304 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.991589 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 14:35:13.304233681 +0000 UTC Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.996118 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.996179 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.996138 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:25 crc kubenswrapper[4836]: E0122 09:43:25.996317 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:25 crc kubenswrapper[4836]: E0122 09:43:25.996426 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:25 crc kubenswrapper[4836]: E0122 09:43:25.996718 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.996807 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:25 crc kubenswrapper[4836]: E0122 09:43:25.996978 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.997579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.997647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.997668 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.997691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:25 crc kubenswrapper[4836]: I0122 09:43:25.997710 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.100369 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.100427 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.100444 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.100463 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.100478 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.203584 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.203637 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.203652 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.203677 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.203693 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.306028 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.306185 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.306205 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.306228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.306244 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.409880 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.409946 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.409963 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.409988 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.410005 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.513765 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.513847 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.513867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.513895 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.513913 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.618250 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.618315 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.618325 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.618346 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.618361 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.720849 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.720912 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.720929 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.720951 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.720967 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.824350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.824475 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.824552 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.824592 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.824665 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.927528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.927816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.927842 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.927870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.927896 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:26 crc kubenswrapper[4836]: I0122 09:43:26.991891 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 14:28:23.347299556 +0000 UTC Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.022923 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2aa3f461-19ed-4700-94b6-8c8dd71c6a21\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eac94552b2f7d92442ea03d2e0437ebcb895ca0a9fbe586ebbb5408ce0a6b43f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d36b29ad56c199e9e3650e0d77eb7142e670d3e8cfb9d813af0c665d2140f6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://770c5287494971d477947ee19d4d7241d8c626ff97d0dcda5d205ecbd06dfc11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0c59d27535afab5577257fef91565a57a111bb61fea65ac804b10f4aa0e8cbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://232268862e531af8eaf2118f4103d61359b7f973bed1cc1245a42c7de6b44548\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac0967a3e5e05e2df23a7ec9ad514672312dd44f4b59b09a7ced5aa6458518db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b25832e70e5cbbca00d6139b91529e2000b361d45ce9d8ac88bb067383e0911\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r26cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gfrrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.031274 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.031493 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.031647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.031797 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.031937 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.047625 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-89st4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fcb3e24-f305-4f39-af85-2dc05a0af79f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"message\\\":\\\"2026-01-22T09:42:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3\\\\n2026-01-22T09:42:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_32502c1c-fc34-4b4e-9e93-b0961b394aa3 to /host/opt/cni/bin/\\\\n2026-01-22T09:42:17Z [verbose] multus-daemon started\\\\n2026-01-22T09:42:17Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2jt5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-89st4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.087214 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc5486e7-b643-4b21-9be6-8d6f475b3908\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:09Z\\\",\\\"message\\\":\\\"eSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9001,TargetPort:{0 9001 },NodePort:0,AppProtocol:nil,},ServicePort{Name:health,Protocol:TCP,Port:8798,TargetPort:{0 8798 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{k8s-app: machine-config-daemon,},ClusterIP:10.217.4.43,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.43],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0122 09:43:08.875072 6834 lb_config.go:1031] Cluster endpoints for openshift-machine-config-operator/machine-config-daemon for network=default are: map[]\\\\nI0122 09:43:08.875072 6834 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-7wkg5\\\\nI0122 09:43:08.875096 6834 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-7wkg5 in node crc\\\\nF0122 09:43:08.875133 6834 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4zbvn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wblv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.109176 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09bc4a6f-04fc-4e7a-bd92-ceac6f457de8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac656ed112ce79d718bfb1cc4725e1449f0d745015dcc1328b43b87304963914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f20f8bfb46e130f3f084b1830c306608dc3c3305590542cc29ad2e0f97892b1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfb9h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6qnv8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.127067 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fe3de3-7ffa-4e0d-b916-95a147ea1b5a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc4da0a8e2fc6052e9b52f6180bac45f8f0b494ae84ef09e7c97d32bd72e70e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcabf7dbae3769a76d1a4ddba12d0aa26af6dbe0f44cb2e426e7142055697d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af0f884a94afa49cf993b12d63e9f37c81c6821cda44f17cf64c0d6f9e205f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c06f2acd0acbe2525c165455e48bb71067b7f5a9ec533139d5fefb72f64f269\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.134490 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.134537 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.134576 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.134597 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.134610 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.140802 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.153731 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a06cebc0-d38a-47ea-a8af-ad38a6471402\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8f2b12dd752755f4fd44a58e2504db7a6177744b507d8dc55e0aab1c8f3e9e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf649233c9b4c938c8b00175f563b7d46947f3e907b8ca43ee83ea9e30b2df1d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3da4b6c19441d8c67457cfc9c379eecb83f755b391fcf4e21f6ad69134b83dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.169499 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f439e8c19f863e13e8f6281f36f88245f9fdba729ac93a1e629f596b9db16b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.183082 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.197253 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8zwn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d7820ac-f475-497d-ac8c-623ef565faee\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce4edf1a4e8c384584c5d669467aaa9aca3e10195ec4408781cd049689a061a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqddz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8zwn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.208785 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"506fe75f-8c00-4768-a50b-80dd7afde450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://26fcf130d5820ff6fd3314a1094a1a590f1cf8dac30d21ce01fd8638cd7a8265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c39c2da321e1366364e44c22754daaf629f0ee4c74c4250be735b4252078020\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.224002 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff57713640f411bc975d05f326d35535cc96d836e8fbe3f3649e8c1b2f37a7ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39a2d1a081b60d309f4259b1821b4676c1a374051534eab85ce3de2c9603125a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.235731 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6adb4a56a6844be66ebb79641ca222cbcfb2fbfb13daae15c89677a9f13f555a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.236608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.236638 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.236646 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.236658 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.236667 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.246713 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.257660 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ea94f0-dd58-4f5b-b61c-b59c074af1f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd33eac3ba2cff98e5b68a6a1ba585f1ae2edc86b22f6a0bb3c91bf1a2a5c38f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-msb79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-n8vsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.265806 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-7wkg5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c414a28-3025-49f7-8213-31f9f2990e97\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99cac06514ebec247d7db1690a7c6c5e871e83e9073a2dee0460d777cee747f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sc8qh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-7wkg5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.274475 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ab725c3-0744-400a-b8d7-66e1165476e5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8jr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6kkmw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.285553 4836 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8da4f34-5379-4b90-8c5b-e33691b4107e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:42:15Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:42:09.388540 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:42:09.389945 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1432337216/tls.crt::/tmp/serving-cert-1432337216/tls.key\\\\\\\"\\\\nI0122 09:42:15.386097 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:42:15.401544 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:42:15.401585 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:42:15.401605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:42:15.401611 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:42:15.406680 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:42:15.406750 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406779 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:42:15.406810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:42:15.406834 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:42:15.406857 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:42:15.406911 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0122 09:42:15.406893 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0122 09:42:15.409405 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:41:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:41:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:41:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:41:57Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.339014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.339121 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.339143 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.339168 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.339187 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.442186 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.442234 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.442245 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.442261 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.442270 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.545409 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.545516 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.545544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.545567 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.545584 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.648196 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.648277 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.648294 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.648341 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.648356 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.751628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.751699 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.751716 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.751740 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.751761 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.854666 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.854715 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.854726 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.854742 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.854754 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.957204 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.957290 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.957302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.957323 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.957336 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.992526 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 15:35:57.13896765 +0000 UTC Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.996883 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.996948 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.996884 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:27 crc kubenswrapper[4836]: I0122 09:43:27.997027 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:27 crc kubenswrapper[4836]: E0122 09:43:27.997184 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:27 crc kubenswrapper[4836]: E0122 09:43:27.997439 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:27 crc kubenswrapper[4836]: E0122 09:43:27.997564 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:27 crc kubenswrapper[4836]: E0122 09:43:27.997647 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.060993 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.061076 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.061088 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.061115 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.061128 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.163874 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.163950 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.163973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.164000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.164021 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.267448 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.267516 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.267533 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.267561 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.267586 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.370419 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.370470 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.370486 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.370509 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.370526 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.472682 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.472732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.472741 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.472755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.472766 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.575930 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.575976 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.575985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.576000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.576009 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.678798 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.678848 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.678859 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.678875 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.678886 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.782393 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.782454 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.782467 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.782489 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.782504 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.885228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.885262 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.885271 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.885285 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.885294 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.988391 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.988453 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.988472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.988497 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.988513 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:28 crc kubenswrapper[4836]: I0122 09:43:28.992763 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 10:34:46.718431461 +0000 UTC Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.090843 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.090911 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.090924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.090938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.090948 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.193754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.193848 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.193867 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.194406 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.194466 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.297393 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.297458 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.297477 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.297518 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.297538 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.401587 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.401651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.401671 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.401698 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.401719 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.503769 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.503818 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.503834 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.503855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.503870 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.606130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.606194 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.606206 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.606232 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.606247 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.708587 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.708664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.708686 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.708713 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.708732 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.811528 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.811600 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.811618 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.811646 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.811664 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.915328 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.915393 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.915420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.915449 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.915470 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.993557 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 12:45:39.608933806 +0000 UTC Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.996388 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.996659 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.996732 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:29 crc kubenswrapper[4836]: I0122 09:43:29.996742 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:29 crc kubenswrapper[4836]: E0122 09:43:29.997114 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:29 crc kubenswrapper[4836]: E0122 09:43:29.997289 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:29 crc kubenswrapper[4836]: E0122 09:43:29.997445 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:29 crc kubenswrapper[4836]: E0122 09:43:29.997518 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.017946 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.018016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.018034 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.018090 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.018107 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.120975 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.121033 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.121094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.121120 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.121141 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.224185 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.224260 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.224279 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.224305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.224323 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.328317 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.328497 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.328529 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.328553 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.328571 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.431665 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.431722 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.431735 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.431752 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.431766 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.534368 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.534410 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.534420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.534435 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.534444 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.637683 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.638155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.638259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.638361 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.638471 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.740469 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.740544 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.740572 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.740598 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.740618 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.844290 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.844329 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.844344 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.844365 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.844380 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.946875 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.946933 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.946947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.946965 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.946979 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:30 crc kubenswrapper[4836]: I0122 09:43:30.994696 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 22:32:02.081927887 +0000 UTC Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.050865 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.050923 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.050939 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.050959 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.050978 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.154514 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.154564 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.154579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.154596 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.154608 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.257374 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.257423 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.257435 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.257454 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.257466 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.359891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.359954 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.359973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.359995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.360012 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.462547 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.462579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.462589 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.462604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.462614 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.565364 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.565429 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.565439 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.565453 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.565461 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.668774 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.668862 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.668885 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.668917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.668940 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.771975 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772053 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772066 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772075 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772738 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772785 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772794 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772803 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.772812 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.794137 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.800028 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.800188 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.800215 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.800247 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.800272 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.823168 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.828608 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.828703 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.828733 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.828767 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.828788 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.848331 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.852850 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.852907 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.852925 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.852948 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.852966 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.872540 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.877775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.877839 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.877863 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.877891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.877913 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.894868 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cfbd5c21-5034-481d-a75e-e8d2d8dd038d\\\",\\\"systemUUID\\\":\\\"ed462382-29a2-4539-9f5b-0dbc69ba727c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z" Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.895105 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.897135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.897168 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.897180 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.897198 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.897209 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.995448 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 19:16:33.977594447 +0000 UTC Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.996817 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.996932 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.996930 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.997030 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.996979 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.997284 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.997353 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:31 crc kubenswrapper[4836]: E0122 09:43:31.997481 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.999025 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.999100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.999119 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.999165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:31 crc kubenswrapper[4836]: I0122 09:43:31.999180 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.101766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.101824 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.101840 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.101865 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.101879 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.205250 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.205330 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.205352 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.205377 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.205395 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.308010 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.308108 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.308134 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.308165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.308189 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.410947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.411004 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.411027 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.411094 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.411118 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.513691 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.513765 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.513784 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.513810 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.513829 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.615980 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.616081 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.616100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.616123 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.616139 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.718603 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.718651 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.718660 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.718676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.718684 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.821302 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.821374 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.821423 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.821449 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.821463 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.923754 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.923829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.923843 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.923866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.923889 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:32 crc kubenswrapper[4836]: I0122 09:43:32.996617 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 10:08:49.70314043 +0000 UTC Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.025724 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.025764 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.025775 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.025793 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.025804 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.128160 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.128201 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.128212 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.128234 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.128250 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.231581 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.231652 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.231674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.231701 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.231722 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.335075 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.335138 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.335155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.335177 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.335194 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.438173 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.438228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.438257 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.438286 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.438308 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.541461 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.541532 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.541555 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.541585 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.541611 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.644884 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.644927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.644947 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.644966 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.644978 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.747075 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.747113 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.747124 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.747142 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.747154 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.850141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.850235 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.850259 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.850287 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.850309 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.953267 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.953320 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.953332 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.953350 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.953366 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.996681 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.996720 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.996739 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.996759 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 01:58:20.538291097 +0000 UTC Jan 22 09:43:33 crc kubenswrapper[4836]: E0122 09:43:33.996806 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:33 crc kubenswrapper[4836]: E0122 09:43:33.996864 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:33 crc kubenswrapper[4836]: I0122 09:43:33.996866 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:33 crc kubenswrapper[4836]: E0122 09:43:33.997127 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:33 crc kubenswrapper[4836]: E0122 09:43:33.997189 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.055100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.055141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.055152 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.055170 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.055180 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.157682 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.157726 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.157736 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.157777 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.157788 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.260797 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.260853 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.260870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.260891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.260908 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.363240 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.363290 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.363299 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.363314 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.363325 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.453648 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:34 crc kubenswrapper[4836]: E0122 09:43:34.453856 4836 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:43:34 crc kubenswrapper[4836]: E0122 09:43:34.454092 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs podName:9ab725c3-0744-400a-b8d7-66e1165476e5 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:38.454008992 +0000 UTC m=+161.813777559 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs") pod "network-metrics-daemon-6kkmw" (UID: "9ab725c3-0744-400a-b8d7-66e1165476e5") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.466689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.466723 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.466784 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.466804 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.466817 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.569635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.569676 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.569688 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.569705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.569718 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.671948 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.671984 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.671995 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.672013 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.672024 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.775415 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.775459 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.775471 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.775486 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.775497 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.878631 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.878685 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.878702 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.878723 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.878738 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.981546 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.981687 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.981710 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.981739 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.981763 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:34 crc kubenswrapper[4836]: I0122 09:43:34.996943 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 15:55:24.423341884 +0000 UTC Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.084587 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.084647 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.084664 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.084690 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.084709 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.187711 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.187788 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.187805 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.187829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.187847 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.290452 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.290520 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.290538 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.290563 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.290581 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.393994 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.394116 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.394146 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.394177 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.394200 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.497090 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.497135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.497151 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.497176 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.497194 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.600244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.600311 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.600329 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.600355 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.600373 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.704515 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.704554 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.704566 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.704583 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.704595 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.807169 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.807244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.807266 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.807298 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.807321 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.910161 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.910223 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.910240 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.910267 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.910288 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.996840 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.996891 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.997126 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:43:24.506278343 +0000 UTC Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.997221 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.997186 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:35 crc kubenswrapper[4836]: E0122 09:43:35.997340 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:35 crc kubenswrapper[4836]: E0122 09:43:35.997434 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:35 crc kubenswrapper[4836]: E0122 09:43:35.997615 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:35 crc kubenswrapper[4836]: E0122 09:43:35.997764 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:35 crc kubenswrapper[4836]: I0122 09:43:35.998730 4836 scope.go:117] "RemoveContainer" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" Jan 22 09:43:35 crc kubenswrapper[4836]: E0122 09:43:35.998979 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-wblv2_openshift-ovn-kubernetes(dc5486e7-b643-4b21-9be6-8d6f475b3908)\"" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.012973 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.013029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.013100 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.013135 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.013162 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.116292 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.116363 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.116387 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.116420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.116447 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.219499 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.219573 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.219597 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.219628 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.219651 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.322666 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.322751 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.322774 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.322806 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.322826 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.425591 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.425674 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.425695 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.425720 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.425736 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.528785 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.528885 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.528908 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.528935 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.528955 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.632870 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.632938 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.632956 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.632981 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.632998 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.735677 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.735747 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.735793 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.735809 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.735821 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.838120 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.838504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.838518 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.838536 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.838547 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.941630 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.941695 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.941718 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.941751 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.941773 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:36 crc kubenswrapper[4836]: I0122 09:43:36.997530 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 11:50:54.03060358 +0000 UTC Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.018990 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.044427 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.044855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.044890 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.044924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.044935 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.060813 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podStartSLOduration=82.060795863 podStartE2EDuration="1m22.060795863s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.06070632 +0000 UTC m=+100.420474837" watchObservedRunningTime="2026-01-22 09:43:37.060795863 +0000 UTC m=+100.420564380" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.069546 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-7wkg5" podStartSLOduration=82.069528371 podStartE2EDuration="1m22.069528371s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.069441739 +0000 UTC m=+100.429210256" watchObservedRunningTime="2026-01-22 09:43:37.069528371 +0000 UTC m=+100.429296888" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.093991 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.093969922 podStartE2EDuration="1m21.093969922s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.093824948 +0000 UTC m=+100.453593465" watchObservedRunningTime="2026-01-22 09:43:37.093969922 +0000 UTC m=+100.453738439" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.121319 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gfrrz" podStartSLOduration=81.121296658 podStartE2EDuration="1m21.121296658s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.121213955 +0000 UTC m=+100.480982493" watchObservedRunningTime="2026-01-22 09:43:37.121296658 +0000 UTC m=+100.481065175" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.147474 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.147738 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.147855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.147965 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.148082 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.159105 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-89st4" podStartSLOduration=81.159086467 podStartE2EDuration="1m21.159086467s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.134503243 +0000 UTC m=+100.494271770" watchObservedRunningTime="2026-01-22 09:43:37.159086467 +0000 UTC m=+100.518854984" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.184234 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.184211245 podStartE2EDuration="52.184211245s" podCreationTimestamp="2026-01-22 09:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.183498427 +0000 UTC m=+100.543266984" watchObservedRunningTime="2026-01-22 09:43:37.184211245 +0000 UTC m=+100.543979762" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.184849 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6qnv8" podStartSLOduration=80.184843371 podStartE2EDuration="1m20.184843371s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.170803694 +0000 UTC m=+100.530572221" watchObservedRunningTime="2026-01-22 09:43:37.184843371 +0000 UTC m=+100.544611888" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.210685 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.210667688 podStartE2EDuration="1m16.210667688s" podCreationTimestamp="2026-01-22 09:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.210484124 +0000 UTC m=+100.570252661" watchObservedRunningTime="2026-01-22 09:43:37.210667688 +0000 UTC m=+100.570436205" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.249749 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.249803 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.249819 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.249840 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.249855 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.250692 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8zwn4" podStartSLOduration=82.250676266 podStartE2EDuration="1m22.250676266s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.250501371 +0000 UTC m=+100.610269888" watchObservedRunningTime="2026-01-22 09:43:37.250676266 +0000 UTC m=+100.610444783" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.272073 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.272025344 podStartE2EDuration="30.272025344s" podCreationTimestamp="2026-01-22 09:43:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:37.260176524 +0000 UTC m=+100.619945041" watchObservedRunningTime="2026-01-22 09:43:37.272025344 +0000 UTC m=+100.631793861" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.352534 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.352600 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.352617 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.352641 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.352660 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.455785 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.455846 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.455863 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.455915 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.455937 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.558414 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.558468 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.558484 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.558504 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.558519 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.661466 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.661579 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.661604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.661675 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.661706 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.765098 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.765158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.765168 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.765185 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.765199 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.867588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.867688 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.867704 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.867720 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.867732 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.970605 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.970665 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.970679 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.970699 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.970717 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.996585 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.996693 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:37 crc kubenswrapper[4836]: E0122 09:43:37.996750 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.996765 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.996785 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:37 crc kubenswrapper[4836]: E0122 09:43:37.996860 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:37 crc kubenswrapper[4836]: E0122 09:43:37.997071 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:37 crc kubenswrapper[4836]: E0122 09:43:37.997260 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:37 crc kubenswrapper[4836]: I0122 09:43:37.998618 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 23:38:25.619420389 +0000 UTC Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.073855 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.073900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.073909 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.073921 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.073929 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.176078 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.176130 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.176142 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.176158 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.176169 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.278924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.278969 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.278992 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.279012 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.279028 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.381927 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.381987 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.382000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.382016 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.382027 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.485263 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.485314 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.485327 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.485346 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.485363 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.587877 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.588167 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.588200 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.588230 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.588252 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.691084 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.691844 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.691888 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.691916 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.691933 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.795404 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.795456 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.795472 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.795494 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.795511 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.897734 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.897807 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.897816 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.897829 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.897838 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:38 crc kubenswrapper[4836]: I0122 09:43:38.998759 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 13:17:06.420608371 +0000 UTC Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.000880 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.000924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.000942 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.001451 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.001465 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.103781 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.103825 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.103836 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.103852 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.103864 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.205904 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.205954 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.205968 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.205986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.205999 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.307817 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.307860 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.307874 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.307891 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.307903 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.411165 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.411225 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.411244 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.411269 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.411281 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.514000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.514090 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.514110 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.514133 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.514149 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.617258 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.617295 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.617305 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.617321 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.617333 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.720937 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.721001 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.721014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.721029 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.721054 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.824635 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.824689 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.824705 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.824726 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.824741 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.927755 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.927851 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.927866 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.927900 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.927913 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.996801 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.996940 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:39 crc kubenswrapper[4836]: E0122 09:43:39.997002 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.997103 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:39 crc kubenswrapper[4836]: E0122 09:43:39.997220 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.997113 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:39 crc kubenswrapper[4836]: E0122 09:43:39.997332 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:39 crc kubenswrapper[4836]: E0122 09:43:39.997464 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:39 crc kubenswrapper[4836]: I0122 09:43:39.999879 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 13:47:35.681777134 +0000 UTC Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.030916 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.030985 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.030997 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.031019 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.031075 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.134141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.134209 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.134221 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.134246 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.134259 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.236594 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.236624 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.236632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.236645 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.236655 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.338393 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.338420 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.338431 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.338443 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.338452 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.440860 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.440917 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.440934 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.440957 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.440977 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.544454 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.544529 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.544540 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.544562 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.544576 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.647262 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.647312 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.647324 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.647339 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.647348 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.750150 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.750211 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.750228 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.750250 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.750266 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.853924 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.853986 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.854018 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.854068 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.854097 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.956878 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.956949 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.956970 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.956997 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:40 crc kubenswrapper[4836]: I0122 09:43:40.957018 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.000965 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:52:41.346056855 +0000 UTC Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.060447 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.060604 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.060632 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.060660 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.060681 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.164014 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.164128 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.164155 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.164182 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.164199 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.267660 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.267716 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.267736 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.267766 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.267784 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.371932 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.372000 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.372011 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.372031 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.372058 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.475529 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.475578 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.475588 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.475605 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.475617 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.578732 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.578804 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.578820 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.578844 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.578861 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.682527 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.682728 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.682780 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.682810 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.682828 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.785290 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.785347 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.785360 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.785379 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.785394 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.887690 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.887733 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.887746 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.887761 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.887771 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.990032 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.990202 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.990226 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.990256 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.990278 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.996637 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.996660 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.996707 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:41 crc kubenswrapper[4836]: E0122 09:43:41.996737 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:41 crc kubenswrapper[4836]: I0122 09:43:41.996773 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:41 crc kubenswrapper[4836]: E0122 09:43:41.996847 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:41 crc kubenswrapper[4836]: E0122 09:43:41.996999 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:41 crc kubenswrapper[4836]: E0122 09:43:41.997190 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.002108 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 23:09:36.587496087 +0000 UTC Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.046966 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.047098 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.047141 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.047173 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.047196 4836 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.107140 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp"] Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.107630 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.109619 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.110268 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.110423 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.112181 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.174139 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.174120228 podStartE2EDuration="5.174120228s" podCreationTimestamp="2026-01-22 09:43:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:42.152617676 +0000 UTC m=+105.512386213" watchObservedRunningTime="2026-01-22 09:43:42.174120228 +0000 UTC m=+105.533888745" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.238484 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.238523 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/167692aa-1804-44c9-9456-b11ffc8a7ee3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.238540 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/167692aa-1804-44c9-9456-b11ffc8a7ee3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.238573 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.238608 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/167692aa-1804-44c9-9456-b11ffc8a7ee3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339641 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/167692aa-1804-44c9-9456-b11ffc8a7ee3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339760 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339813 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/167692aa-1804-44c9-9456-b11ffc8a7ee3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339857 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/167692aa-1804-44c9-9456-b11ffc8a7ee3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339876 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.339966 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.340117 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/167692aa-1804-44c9-9456-b11ffc8a7ee3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.340783 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/167692aa-1804-44c9-9456-b11ffc8a7ee3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.346964 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/167692aa-1804-44c9-9456-b11ffc8a7ee3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.357310 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/167692aa-1804-44c9-9456-b11ffc8a7ee3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zq4kp\" (UID: \"167692aa-1804-44c9-9456-b11ffc8a7ee3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.425600 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" Jan 22 09:43:42 crc kubenswrapper[4836]: W0122 09:43:42.446108 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod167692aa_1804_44c9_9456_b11ffc8a7ee3.slice/crio-9f02bcf4d56d5de5086bc86ffc061aacfab8858245633ce7e4e23de1d51f991b WatchSource:0}: Error finding container 9f02bcf4d56d5de5086bc86ffc061aacfab8858245633ce7e4e23de1d51f991b: Status 404 returned error can't find the container with id 9f02bcf4d56d5de5086bc86ffc061aacfab8858245633ce7e4e23de1d51f991b Jan 22 09:43:42 crc kubenswrapper[4836]: I0122 09:43:42.689314 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" event={"ID":"167692aa-1804-44c9-9456-b11ffc8a7ee3","Type":"ContainerStarted","Data":"9f02bcf4d56d5de5086bc86ffc061aacfab8858245633ce7e4e23de1d51f991b"} Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.002381 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 14:02:34.115061111 +0000 UTC Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.002517 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.014826 4836 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.692519 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" event={"ID":"167692aa-1804-44c9-9456-b11ffc8a7ee3","Type":"ContainerStarted","Data":"2ddcc8c4b27f3db6e9ffb9ea4170343dc9be8b3720d18c015bb884228e0ce3df"} Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.706933 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zq4kp" podStartSLOduration=87.706916963 podStartE2EDuration="1m27.706916963s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:43.705931026 +0000 UTC m=+107.065699553" watchObservedRunningTime="2026-01-22 09:43:43.706916963 +0000 UTC m=+107.066685490" Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.996401 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.996436 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.996415 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:43 crc kubenswrapper[4836]: I0122 09:43:43.996603 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:43 crc kubenswrapper[4836]: E0122 09:43:43.996593 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:43 crc kubenswrapper[4836]: E0122 09:43:43.996709 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:43 crc kubenswrapper[4836]: E0122 09:43:43.996807 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:43 crc kubenswrapper[4836]: E0122 09:43:43.996931 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:45 crc kubenswrapper[4836]: I0122 09:43:45.997010 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:45 crc kubenswrapper[4836]: I0122 09:43:45.997128 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:45 crc kubenswrapper[4836]: I0122 09:43:45.997053 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:45 crc kubenswrapper[4836]: I0122 09:43:45.997181 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:45 crc kubenswrapper[4836]: E0122 09:43:45.997242 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:45 crc kubenswrapper[4836]: E0122 09:43:45.997396 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:45 crc kubenswrapper[4836]: E0122 09:43:45.997477 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:45 crc kubenswrapper[4836]: E0122 09:43:45.997571 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:47 crc kubenswrapper[4836]: I0122 09:43:47.996925 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:47 crc kubenswrapper[4836]: E0122 09:43:47.997783 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:47 crc kubenswrapper[4836]: I0122 09:43:47.997090 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:47 crc kubenswrapper[4836]: I0122 09:43:47.997072 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:47 crc kubenswrapper[4836]: E0122 09:43:47.997906 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:47 crc kubenswrapper[4836]: I0122 09:43:47.997159 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:47 crc kubenswrapper[4836]: E0122 09:43:47.998099 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:47 crc kubenswrapper[4836]: E0122 09:43:47.998282 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:49 crc kubenswrapper[4836]: I0122 09:43:49.996389 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:49 crc kubenswrapper[4836]: I0122 09:43:49.996414 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:49 crc kubenswrapper[4836]: I0122 09:43:49.996464 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:49 crc kubenswrapper[4836]: I0122 09:43:49.996508 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:49 crc kubenswrapper[4836]: E0122 09:43:49.997301 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:49 crc kubenswrapper[4836]: E0122 09:43:49.997540 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:49 crc kubenswrapper[4836]: I0122 09:43:49.997635 4836 scope.go:117] "RemoveContainer" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" Jan 22 09:43:49 crc kubenswrapper[4836]: E0122 09:43:49.997649 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:49 crc kubenswrapper[4836]: E0122 09:43:49.997419 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.717618 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/1.log" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.718122 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/0.log" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.718165 4836 generic.go:334] "Generic (PLEG): container finished" podID="7fcb3e24-f305-4f39-af85-2dc05a0af79f" containerID="023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633" exitCode=1 Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.718227 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerDied","Data":"023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633"} Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.718264 4836 scope.go:117] "RemoveContainer" containerID="07091e0272190cf72f93dedc57926ef185d1010ffa855bdeef299d7e5caab784" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.718652 4836 scope.go:117] "RemoveContainer" containerID="023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633" Jan 22 09:43:50 crc kubenswrapper[4836]: E0122 09:43:50.718787 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-89st4_openshift-multus(7fcb3e24-f305-4f39-af85-2dc05a0af79f)\"" pod="openshift-multus/multus-89st4" podUID="7fcb3e24-f305-4f39-af85-2dc05a0af79f" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.724289 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/3.log" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.727026 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerStarted","Data":"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba"} Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.727397 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:43:50 crc kubenswrapper[4836]: I0122 09:43:50.770229 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podStartSLOduration=94.770186588 podStartE2EDuration="1m34.770186588s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:43:50.769993553 +0000 UTC m=+114.129762070" watchObservedRunningTime="2026-01-22 09:43:50.770186588 +0000 UTC m=+114.129955125" Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.054265 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6kkmw"] Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.054406 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:51 crc kubenswrapper[4836]: E0122 09:43:51.054509 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.731226 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/1.log" Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.996251 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.996278 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:51 crc kubenswrapper[4836]: I0122 09:43:51.996258 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:51 crc kubenswrapper[4836]: E0122 09:43:51.996408 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:51 crc kubenswrapper[4836]: E0122 09:43:51.996642 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:51 crc kubenswrapper[4836]: E0122 09:43:51.996839 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:52 crc kubenswrapper[4836]: I0122 09:43:52.997097 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:52 crc kubenswrapper[4836]: E0122 09:43:52.997253 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:53 crc kubenswrapper[4836]: I0122 09:43:53.996520 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:53 crc kubenswrapper[4836]: I0122 09:43:53.996631 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:53 crc kubenswrapper[4836]: I0122 09:43:53.996520 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:53 crc kubenswrapper[4836]: E0122 09:43:53.996730 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:53 crc kubenswrapper[4836]: E0122 09:43:53.996922 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:53 crc kubenswrapper[4836]: E0122 09:43:53.997117 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:54 crc kubenswrapper[4836]: I0122 09:43:54.997203 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:54 crc kubenswrapper[4836]: E0122 09:43:54.997427 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:55 crc kubenswrapper[4836]: I0122 09:43:55.996500 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:55 crc kubenswrapper[4836]: E0122 09:43:55.996642 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:55 crc kubenswrapper[4836]: I0122 09:43:55.996502 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:55 crc kubenswrapper[4836]: I0122 09:43:55.996500 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:55 crc kubenswrapper[4836]: E0122 09:43:55.996908 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:55 crc kubenswrapper[4836]: E0122 09:43:55.997092 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:56 crc kubenswrapper[4836]: E0122 09:43:56.942902 4836 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 22 09:43:56 crc kubenswrapper[4836]: I0122 09:43:56.996683 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:56 crc kubenswrapper[4836]: E0122 09:43:56.999420 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:57 crc kubenswrapper[4836]: E0122 09:43:57.104758 4836 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 09:43:57 crc kubenswrapper[4836]: I0122 09:43:57.996233 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:58 crc kubenswrapper[4836]: I0122 09:43:57.996223 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:58 crc kubenswrapper[4836]: E0122 09:43:57.996437 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:43:58 crc kubenswrapper[4836]: I0122 09:43:57.996245 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:58 crc kubenswrapper[4836]: E0122 09:43:57.996582 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:58 crc kubenswrapper[4836]: E0122 09:43:57.996760 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:58 crc kubenswrapper[4836]: I0122 09:43:58.996586 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:43:58 crc kubenswrapper[4836]: E0122 09:43:58.996792 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:43:59 crc kubenswrapper[4836]: I0122 09:43:59.996459 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:43:59 crc kubenswrapper[4836]: I0122 09:43:59.996459 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:43:59 crc kubenswrapper[4836]: E0122 09:43:59.997209 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:43:59 crc kubenswrapper[4836]: E0122 09:43:59.997345 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:43:59 crc kubenswrapper[4836]: I0122 09:43:59.996498 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:43:59 crc kubenswrapper[4836]: E0122 09:43:59.997460 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:44:00 crc kubenswrapper[4836]: I0122 09:44:00.996591 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:00 crc kubenswrapper[4836]: E0122 09:44:00.996884 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:44:00 crc kubenswrapper[4836]: I0122 09:44:00.997114 4836 scope.go:117] "RemoveContainer" containerID="023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633" Jan 22 09:44:01 crc kubenswrapper[4836]: I0122 09:44:01.996498 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:01 crc kubenswrapper[4836]: I0122 09:44:01.996796 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:01 crc kubenswrapper[4836]: I0122 09:44:01.996840 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:01 crc kubenswrapper[4836]: E0122 09:44:01.997178 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:44:01 crc kubenswrapper[4836]: E0122 09:44:01.997270 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:44:01 crc kubenswrapper[4836]: E0122 09:44:01.997336 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:44:02 crc kubenswrapper[4836]: E0122 09:44:02.105814 4836 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 09:44:02 crc kubenswrapper[4836]: I0122 09:44:02.769958 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/1.log" Jan 22 09:44:02 crc kubenswrapper[4836]: I0122 09:44:02.770030 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerStarted","Data":"7ce729b6cbd16f82ef39e221e8774e4f11535f85358b4a0f4e713f3221ec39e9"} Jan 22 09:44:02 crc kubenswrapper[4836]: I0122 09:44:02.996374 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:02 crc kubenswrapper[4836]: E0122 09:44:02.996553 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:44:03 crc kubenswrapper[4836]: I0122 09:44:03.996465 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:03 crc kubenswrapper[4836]: E0122 09:44:03.996604 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:44:03 crc kubenswrapper[4836]: I0122 09:44:03.996810 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:03 crc kubenswrapper[4836]: E0122 09:44:03.996874 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:44:03 crc kubenswrapper[4836]: I0122 09:44:03.997011 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:03 crc kubenswrapper[4836]: E0122 09:44:03.997113 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:44:04 crc kubenswrapper[4836]: I0122 09:44:04.996949 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:04 crc kubenswrapper[4836]: E0122 09:44:04.997077 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:44:05 crc kubenswrapper[4836]: I0122 09:44:05.996416 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:05 crc kubenswrapper[4836]: I0122 09:44:05.996491 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:05 crc kubenswrapper[4836]: E0122 09:44:05.996552 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 22 09:44:05 crc kubenswrapper[4836]: E0122 09:44:05.996637 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 22 09:44:05 crc kubenswrapper[4836]: I0122 09:44:05.996491 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:05 crc kubenswrapper[4836]: E0122 09:44:05.996778 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 22 09:44:06 crc kubenswrapper[4836]: I0122 09:44:06.996402 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:06 crc kubenswrapper[4836]: E0122 09:44:06.997301 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6kkmw" podUID="9ab725c3-0744-400a-b8d7-66e1165476e5" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.996436 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.996487 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.996487 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.998557 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.998725 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.998845 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 09:44:07 crc kubenswrapper[4836]: I0122 09:44:07.999878 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 09:44:08 crc kubenswrapper[4836]: I0122 09:44:08.996826 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:08 crc kubenswrapper[4836]: I0122 09:44:08.999513 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 09:44:09 crc kubenswrapper[4836]: I0122 09:44:08.999979 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.702806 4836 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.740473 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.741457 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.744303 4836 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.744357 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.744380 4836 reflector.go:561] object-"openshift-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.744445 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.744472 4836 reflector.go:561] object-"openshift-controller-manager"/"openshift-global-ca": failed to list *v1.ConfigMap: configmaps "openshift-global-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.744519 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-global-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-global-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.743944 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-txqw8"] Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.745081 4836 reflector.go:561] object-"openshift-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.745244 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.745416 4836 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.745488 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.745517 4836 reflector.go:561] object-"openshift-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.745559 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.745730 4836 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.745771 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.746314 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.749832 4836 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.750100 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.749872 4836 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.750383 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.750465 4836 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.750639 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.751184 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.751627 4836 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.751785 4836 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.751843 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.751792 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.751947 4836 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.751980 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.752143 4836 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.752188 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.751724 4836 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.752283 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.752511 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 09:44:12 crc kubenswrapper[4836]: W0122 09:44:12.752657 4836 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Jan 22 09:44:12 crc kubenswrapper[4836]: E0122 09:44:12.752803 4836 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.755172 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btclw"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.757699 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.758028 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.758027 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.761631 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.762389 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.763195 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.763661 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.763784 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.764314 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.763783 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.764914 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.765398 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.769465 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.769566 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.769585 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.769832 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.769956 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.770008 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.771164 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.771859 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773489 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773542 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773571 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773604 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwwwj\" (UniqueName: \"kubernetes.io/projected/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-kube-api-access-pwwwj\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773635 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773661 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-images\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773691 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-node-pullsecrets\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773718 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fglz\" (UniqueName: \"kubernetes.io/projected/42863163-25f6-41d7-953b-804ba3e052c8-kube-api-access-4fglz\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773746 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773773 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-client\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773800 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8wj\" (UniqueName: \"kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773846 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773872 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/42863163-25f6-41d7-953b-804ba3e052c8-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.773902 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774096 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-audit-dir\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774176 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-encryption-config\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774229 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774284 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774308 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774325 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774369 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774385 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774400 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57ck8\" (UniqueName: \"kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774450 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-config\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774465 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-audit-policies\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774515 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774530 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwd94\" (UniqueName: \"kubernetes.io/projected/172b6aec-4183-447f-82d0-5904cbbc7604-kube-api-access-pwd94\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774547 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774598 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774618 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-serving-cert\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774674 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774689 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774745 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774768 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dqd\" (UniqueName: \"kubernetes.io/projected/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-kube-api-access-74dqd\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774809 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774843 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774880 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774899 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774915 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1121b7db-f2c3-4900-81c7-e765a19a091c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774929 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm6d5\" (UniqueName: \"kubernetes.io/projected/1121b7db-f2c3-4900-81c7-e765a19a091c-kube-api-access-lm6d5\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774952 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774966 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774981 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.774996 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775009 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775024 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/172b6aec-4183-447f-82d0-5904cbbc7604-audit-dir\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775076 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775093 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775110 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775128 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775144 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775270 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.775593 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.776790 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-z9w5n"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.777424 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.778788 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b9fsf"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.779322 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.785150 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.785568 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.785866 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.786833 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.787460 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.787562 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.787769 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.787807 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788000 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788214 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788339 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788421 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788693 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788826 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.788928 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789026 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789337 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789470 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789574 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789692 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789826 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.789922 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.790022 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.816918 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.817089 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.817219 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.817391 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.817736 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.817804 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.818821 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.819478 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.819681 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.819826 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-88xxc"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.820237 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.820463 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.820774 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wp857"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821158 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821289 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821443 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821630 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821720 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.821934 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.822457 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.822694 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.822800 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823179 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823353 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823447 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823536 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823637 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823738 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823862 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823873 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.823883 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.824100 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.824451 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.825004 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.825381 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.826071 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.826475 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.826859 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.826975 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.827219 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.827339 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.827653 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.831555 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.831711 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.831797 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.831871 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.835015 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.835309 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.835468 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.835634 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.835804 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.836013 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.836640 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.837013 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.838003 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.838327 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.840298 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.840531 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.840694 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.840798 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9mq8w"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.841324 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.841651 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.841700 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.841944 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.842475 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.843210 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.843890 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.847132 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.847286 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.853587 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.855124 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.871679 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.871689 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878136 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878180 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878197 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czlcn\" (UniqueName: \"kubernetes.io/projected/ca83435d-4c56-403e-90a9-3afd1fbd417d-kube-api-access-czlcn\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878217 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878233 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878247 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1121b7db-f2c3-4900-81c7-e765a19a091c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878261 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm6d5\" (UniqueName: \"kubernetes.io/projected/1121b7db-f2c3-4900-81c7-e765a19a091c-kube-api-access-lm6d5\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878285 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878301 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878317 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2n8d\" (UniqueName: \"kubernetes.io/projected/f517dfdc-296c-41c3-8506-dc95aa8504db-kube-api-access-z2n8d\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878332 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878348 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878362 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878376 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/172b6aec-4183-447f-82d0-5904cbbc7604-audit-dir\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878389 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f517dfdc-296c-41c3-8506-dc95aa8504db-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878406 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878421 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878437 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878452 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878466 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878482 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a588b581-1bb9-4c23-8be2-26816172ddeb-config\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878498 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a588b581-1bb9-4c23-8be2-26816172ddeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878521 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878536 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878553 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878569 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwwwj\" (UniqueName: \"kubernetes.io/projected/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-kube-api-access-pwwwj\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878604 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878597 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pw59n"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878620 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-images\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878637 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f517dfdc-296c-41c3-8506-dc95aa8504db-serving-cert\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878652 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878669 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-node-pullsecrets\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878688 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fglz\" (UniqueName: \"kubernetes.io/projected/42863163-25f6-41d7-953b-804ba3e052c8-kube-api-access-4fglz\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878703 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878718 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-client\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878732 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878749 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8wj\" (UniqueName: \"kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878763 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a588b581-1bb9-4c23-8be2-26816172ddeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878787 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878803 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/42863163-25f6-41d7-953b-804ba3e052c8-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878821 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878836 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878858 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-audit-dir\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878873 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-encryption-config\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878887 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca83435d-4c56-403e-90a9-3afd1fbd417d-proxy-tls\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878904 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878920 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878935 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-images\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878952 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878969 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878987 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.879001 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.879025 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.879059 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.879597 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880168 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880370 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.879062 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880567 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57ck8\" (UniqueName: \"kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880589 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-config\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880612 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-audit-policies\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880641 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880663 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880686 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880706 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwd94\" (UniqueName: \"kubernetes.io/projected/172b6aec-4183-447f-82d0-5904cbbc7604-kube-api-access-pwd94\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880730 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880754 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880775 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5vq2\" (UniqueName: \"kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880794 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880814 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-serving-cert\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880836 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczsh\" (UniqueName: \"kubernetes.io/projected/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-kube-api-access-jczsh\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880855 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880875 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880907 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880926 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74dqd\" (UniqueName: \"kubernetes.io/projected/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-kube-api-access-74dqd\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.880948 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.878160 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.881776 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.882841 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-config\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.883325 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-audit-policies\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.883484 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.883914 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.884825 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1121b7db-f2c3-4900-81c7-e765a19a091c-images\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.886213 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.886740 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-node-pullsecrets\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.886957 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887223 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887268 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887321 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/172b6aec-4183-447f-82d0-5904cbbc7604-audit-dir\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887662 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-auth-proxy-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887815 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btclw"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.887838 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888181 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888405 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888424 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/09d81482-c475-480e-ad24-1e5eee403a99-audit-dir\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888430 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-serving-cert\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888614 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42863163-25f6-41d7-953b-804ba3e052c8-config\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888911 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.888910 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.889256 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/172b6aec-4183-447f-82d0-5904cbbc7604-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.889359 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.890164 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.890637 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.890684 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.891116 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.891661 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.892171 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g5pts"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.892793 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.892954 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/42863163-25f6-41d7-953b-804ba3e052c8-machine-approver-tls\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.893117 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.893380 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.894677 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-etcd-client\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.895078 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.898457 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.899160 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.899477 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlfc7"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.899536 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.899834 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.899867 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.900836 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.903113 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.903892 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.904293 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.904529 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.905716 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.907668 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.907805 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1121b7db-f2c3-4900-81c7-e765a19a091c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.907858 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.908070 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.908242 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.909216 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.909998 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.910126 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.910192 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.910702 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.911772 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/172b6aec-4183-447f-82d0-5904cbbc7604-encryption-config\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.915869 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.916546 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.916621 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.917566 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.918903 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-txqw8"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.919871 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.920438 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.921296 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.923131 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s7pq7"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.924199 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.924299 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.925354 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.926676 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.930820 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.934416 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wp857"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.935879 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z9w5n"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.937696 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.939055 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.940339 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pw59n"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.941432 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.944079 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.958050 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.969530 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.970258 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.971336 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.972537 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.973604 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b9fsf"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.974635 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.975970 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.977332 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-88xxc"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.978623 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.979923 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.981300 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.981860 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca83435d-4c56-403e-90a9-3afd1fbd417d-proxy-tls\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982000 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-images\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982121 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982238 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982331 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982451 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982561 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5vq2\" (UniqueName: \"kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982663 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982770 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczsh\" (UniqueName: \"kubernetes.io/projected/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-kube-api-access-jczsh\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.982879 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983089 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czlcn\" (UniqueName: \"kubernetes.io/projected/ca83435d-4c56-403e-90a9-3afd1fbd417d-kube-api-access-czlcn\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983211 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983314 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983469 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2n8d\" (UniqueName: \"kubernetes.io/projected/f517dfdc-296c-41c3-8506-dc95aa8504db-kube-api-access-z2n8d\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983891 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f517dfdc-296c-41c3-8506-dc95aa8504db-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984141 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a588b581-1bb9-4c23-8be2-26816172ddeb-config\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984010 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984201 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.983294 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984264 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a588b581-1bb9-4c23-8be2-26816172ddeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984445 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984479 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f517dfdc-296c-41c3-8506-dc95aa8504db-serving-cert\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984481 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f517dfdc-296c-41c3-8506-dc95aa8504db-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984950 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.984978 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a588b581-1bb9-4c23-8be2-26816172ddeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.985380 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.985384 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.985949 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-c7hkn"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.985964 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.986613 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.986701 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.986719 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.987550 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f517dfdc-296c-41c3-8506-dc95aa8504db-serving-cert\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.987599 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.987632 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lll5j"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.988181 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.989315 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.990821 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.992157 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlfc7"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.993476 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c7hkn"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.995048 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r"] Jan 22 09:44:12 crc kubenswrapper[4836]: I0122 09:44:12.996390 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.001496 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g5pts"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.001521 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.001532 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.003349 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.004283 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.006445 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.007922 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a588b581-1bb9-4c23-8be2-26816172ddeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.008413 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s7pq7"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.009871 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hdgd6"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.010450 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.011365 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hdgd6"] Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.025210 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.043428 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.045811 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a588b581-1bb9-4c23-8be2-26816172ddeb-config\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.063895 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.073895 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.083312 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.103482 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.123105 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.143242 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.163523 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.183592 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.205186 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.223903 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.250540 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.267143 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.283308 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.304279 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.333624 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.344149 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.364540 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.384259 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.393842 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ca83435d-4c56-403e-90a9-3afd1fbd417d-images\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.406001 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.415405 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ca83435d-4c56-403e-90a9-3afd1fbd417d-proxy-tls\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.425212 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.463654 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.485323 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.505307 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.525220 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.544980 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.564825 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.584987 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.604008 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.632859 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.644467 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.663844 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.700688 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57ck8\" (UniqueName: \"kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8\") pod \"oauth-openshift-558db77b4-jwkz2\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.737259 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwd94\" (UniqueName: \"kubernetes.io/projected/172b6aec-4183-447f-82d0-5904cbbc7604-kube-api-access-pwd94\") pod \"apiserver-7bbb656c7d-lsxgt\" (UID: \"172b6aec-4183-447f-82d0-5904cbbc7604\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.768329 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dqd\" (UniqueName: \"kubernetes.io/projected/de54019b-ce0c-4dd3-8c94-121cf6d0d5ec-kube-api-access-74dqd\") pod \"cluster-samples-operator-665b6dd947-pn4sd\" (UID: \"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.779033 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm6d5\" (UniqueName: \"kubernetes.io/projected/1121b7db-f2c3-4900-81c7-e765a19a091c-kube-api-access-lm6d5\") pod \"machine-api-operator-5694c8668f-btclw\" (UID: \"1121b7db-f2c3-4900-81c7-e765a19a091c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.782804 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.801816 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fglz\" (UniqueName: \"kubernetes.io/projected/42863163-25f6-41d7-953b-804ba3e052c8-kube-api-access-4fglz\") pod \"machine-approver-56656f9798-fkgzv\" (UID: \"42863163-25f6-41d7-953b-804ba3e052c8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.823511 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.839539 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.843507 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.847156 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.864891 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.884102 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886194 4836 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886293 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.386271464 +0000 UTC m=+137.746039991 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886612 4836 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886655 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.386641634 +0000 UTC m=+137.746410161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886662 4836 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886677 4836 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886703 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.386695116 +0000 UTC m=+137.746463643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886711 4836 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886718 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.386711006 +0000 UTC m=+137.746479543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.886743 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.386727387 +0000 UTC m=+137.746495904 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888215 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888284 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.38826726 +0000 UTC m=+137.748035787 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888320 4836 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888344 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.388337262 +0000 UTC m=+137.748105789 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888371 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888401 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.388391673 +0000 UTC m=+137.748160300 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888550 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888613 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888643 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.38862129 +0000 UTC m=+137.748389807 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888666 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.388655851 +0000 UTC m=+137.748424488 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888717 4836 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.888760 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.388750443 +0000 UTC m=+137.748518970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync secret cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.889866 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: E0122 09:44:13.889993 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:14.389979399 +0000 UTC m=+137.749747916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.902532 4836 request.go:700] Waited for 1.013446635s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager-operator/serviceaccounts/openshift-controller-manager-operator/token Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.924809 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwwwj\" (UniqueName: \"kubernetes.io/projected/dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a-kube-api-access-pwwwj\") pod \"openshift-controller-manager-operator-756b6f6bc6-6jmp5\" (UID: \"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.943503 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.963934 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.983838 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.991256 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" Jan 22 09:44:13 crc kubenswrapper[4836]: I0122 09:44:13.996816 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd"] Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.004511 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.023898 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.027279 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.038236 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.044249 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.044536 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.048870 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt"] Jan 22 09:44:14 crc kubenswrapper[4836]: W0122 09:44:14.050763 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cbc633b_ae56_438d_88c3_91437b545f38.slice/crio-d611aca7aa9f6a4396c0976f15161ea4233162f9d4ab35a527641b65cba09e0e WatchSource:0}: Error finding container d611aca7aa9f6a4396c0976f15161ea4233162f9d4ab35a527641b65cba09e0e: Status 404 returned error can't find the container with id d611aca7aa9f6a4396c0976f15161ea4233162f9d4ab35a527641b65cba09e0e Jan 22 09:44:14 crc kubenswrapper[4836]: W0122 09:44:14.060376 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod172b6aec_4183_447f_82d0_5904cbbc7604.slice/crio-646eb78d7977630baf2a390917cd9c199a8ad15f29b4dba8c1c4262ecb167295 WatchSource:0}: Error finding container 646eb78d7977630baf2a390917cd9c199a8ad15f29b4dba8c1c4262ecb167295: Status 404 returned error can't find the container with id 646eb78d7977630baf2a390917cd9c199a8ad15f29b4dba8c1c4262ecb167295 Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.064305 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.084878 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.104180 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.123810 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.143625 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.163823 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.165774 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5"] Jan 22 09:44:14 crc kubenswrapper[4836]: W0122 09:44:14.172322 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc87ab7f_7d2d_48da_bb1a_a5905f8cb40a.slice/crio-1677464c5e53da8db771f9d0209f693991c0d0556514ff27f28ab5333d6bd7ac WatchSource:0}: Error finding container 1677464c5e53da8db771f9d0209f693991c0d0556514ff27f28ab5333d6bd7ac: Status 404 returned error can't find the container with id 1677464c5e53da8db771f9d0209f693991c0d0556514ff27f28ab5333d6bd7ac Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.185739 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.203843 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.211674 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btclw"] Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.224801 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.243761 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.263430 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.283650 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.303273 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.323455 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.342986 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.363815 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.383543 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.404393 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408320 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408370 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408407 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408442 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408501 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408542 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408585 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408619 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408654 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408682 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408721 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.408756 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.424972 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.445230 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.474361 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.484260 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.504262 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.523167 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.543845 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.565006 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.583738 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.604147 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.623964 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.644358 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.663724 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.684760 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.705081 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.725217 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.744150 4836 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.764250 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.805789 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czlcn\" (UniqueName: \"kubernetes.io/projected/ca83435d-4c56-403e-90a9-3afd1fbd417d-kube-api-access-czlcn\") pod \"machine-config-operator-74547568cd-g5kkn\" (UID: \"ca83435d-4c56-403e-90a9-3afd1fbd417d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:14 crc kubenswrapper[4836]: E0122 09:44:14.814085 4836 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.820525 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczsh\" (UniqueName: \"kubernetes.io/projected/ac27fcac-0bc1-41a1-aac0-2fdce1d06c04-kube-api-access-jczsh\") pod \"kube-storage-version-migrator-operator-b67b599dd-frpzv\" (UID: \"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.834931 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" event={"ID":"42863163-25f6-41d7-953b-804ba3e052c8","Type":"ContainerStarted","Data":"0512e38fead305951213b88bb8943975749b9425acaf72da601bd4fa2ea9915a"} Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.836374 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" event={"ID":"4cbc633b-ae56-438d-88c3-91437b545f38","Type":"ContainerStarted","Data":"d611aca7aa9f6a4396c0976f15161ea4233162f9d4ab35a527641b65cba09e0e"} Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.838282 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" event={"ID":"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a","Type":"ContainerStarted","Data":"1677464c5e53da8db771f9d0209f693991c0d0556514ff27f28ab5333d6bd7ac"} Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.839559 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" event={"ID":"172b6aec-4183-447f-82d0-5904cbbc7604","Type":"ContainerStarted","Data":"646eb78d7977630baf2a390917cd9c199a8ad15f29b4dba8c1c4262ecb167295"} Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.846003 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.856538 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2n8d\" (UniqueName: \"kubernetes.io/projected/f517dfdc-296c-41c3-8506-dc95aa8504db-kube-api-access-z2n8d\") pod \"openshift-config-operator-7777fb866f-vd5ft\" (UID: \"f517dfdc-296c-41c3-8506-dc95aa8504db\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.861494 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5vq2\" (UniqueName: \"kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2\") pod \"console-f9d7485db-tbm5x\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.874657 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.880748 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a588b581-1bb9-4c23-8be2-26816172ddeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qvxjj\" (UID: \"a588b581-1bb9-4c23-8be2-26816172ddeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.883565 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.903645 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.922233 4836 request.go:700] Waited for 1.935280558s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.923705 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 09:44:14 crc kubenswrapper[4836]: E0122 09:44:14.937003 4836 projected.go:288] Couldn't get configMap openshift-apiserver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:14 crc kubenswrapper[4836]: E0122 09:44:14.937064 4836 projected.go:194] Error preparing data for projected volume kube-api-access-xg8wj for pod openshift-apiserver/apiserver-76f77b778f-txqw8: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:14 crc kubenswrapper[4836]: E0122 09:44:14.937142 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.43711578 +0000 UTC m=+138.796884307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xg8wj" (UniqueName: "kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.943162 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.964861 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 09:44:14 crc kubenswrapper[4836]: I0122 09:44:14.984940 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.004284 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.024366 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.044167 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.062213 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.063782 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.079195 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.105093 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.110138 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.135971 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.141580 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.145834 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac99467a-7988-41f3-96bf-908859713734-serving-cert\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.145901 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146238 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-config\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146625 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qjhs\" (UniqueName: \"kubernetes.io/projected/ac99467a-7988-41f3-96bf-908859713734-kube-api-access-5qjhs\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146720 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-trusted-ca\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146816 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146906 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.146990 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147033 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147103 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krspr\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147587 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147670 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/528c6e94-51ea-42a2-bb57-f521453e46bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147707 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528c6e94-51ea-42a2-bb57-f521453e46bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147837 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147881 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147915 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4klnh\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-kube-api-access-4klnh\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.147948 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ggf\" (UniqueName: \"kubernetes.io/projected/c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45-kube-api-access-g6ggf\") pod \"downloads-7954f5f757-z9w5n\" (UID: \"c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45\") " pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.147976 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.64795604 +0000 UTC m=+139.007724587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.151350 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.165422 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.170453 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.185189 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.205191 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.210203 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-image-import-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.228947 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.232761 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-etcd-client\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.243402 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.248880 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249013 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528c6e94-51ea-42a2-bb57-f521453e46bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249087 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249114 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ggf\" (UniqueName: \"kubernetes.io/projected/c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45-kube-api-access-g6ggf\") pod \"downloads-7954f5f757-z9w5n\" (UID: \"c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45\") " pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249143 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stmhf\" (UniqueName: \"kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249167 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.249203 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.749182876 +0000 UTC m=+139.108951393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249231 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x646h\" (UniqueName: \"kubernetes.io/projected/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-kube-api-access-x646h\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249311 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92cea25b-5193-412c-bd80-eb4932e7a7a5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249411 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-etcd-serving-ca\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249440 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28391950-7de5-486a-8394-5a96de836f07-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249498 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249515 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac99467a-7988-41f3-96bf-908859713734-serving-cert\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249535 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-config\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249549 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb0c14a-689b-494a-80b8-ce1df08a29e5-serving-cert\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249585 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-config\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249601 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n89t5\" (UniqueName: \"kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249631 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249666 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249682 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8zkk\" (UniqueName: \"kubernetes.io/projected/70bd79c6-78d6-4da2-8868-40694cb22fe6-kube-api-access-w8zkk\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249698 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-webhook-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249714 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-proxy-tls\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249729 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdr8h\" (UniqueName: \"kubernetes.io/projected/6a3ebaf3-caf9-43e8-aa07-221af07bff44-kube-api-access-vdr8h\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249743 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249760 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqb4v\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-kube-api-access-sqb4v\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249777 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249790 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc211904-b2bb-4dc9-90cb-4e30abdd981f-cert\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249817 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcgtk\" (UniqueName: \"kubernetes.io/projected/368ef53c-70f8-4c26-b844-b45a07cc28eb-kube-api-access-bcgtk\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249834 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/abca25a5-e96f-4e09-8a43-ab855b44d7a2-tmpfs\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249847 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-certs\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249874 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-serving-cert\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249908 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-socket-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.249992 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70bd79c6-78d6-4da2-8868-40694cb22fe6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250007 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e4ea6-63ca-4b08-9386-995aa6a67b5a-serving-cert\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250024 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-csi-data-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250061 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250081 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250100 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250163 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250181 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250220 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd8a6658-4df0-49a1-8853-a038631b2bf3-metrics-tls\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250255 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krspr\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250277 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250302 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250337 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrc6h\" (UniqueName: \"kubernetes.io/projected/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-kube-api-access-xrc6h\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250954 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-srv-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250979 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.251003 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.250980 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.251107 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.251143 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.251187 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5q8r\" (UniqueName: \"kubernetes.io/projected/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-kube-api-access-x5q8r\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.251560 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-config\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.252108 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.254708 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/528c6e94-51ea-42a2-bb57-f521453e46bf-metrics-tls\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.255326 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.255922 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/528c6e94-51ea-42a2-bb57-f521453e46bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.255970 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256015 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-stats-auth\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256059 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx9bx\" (UniqueName: \"kubernetes.io/projected/09809668-72b4-4a16-a821-3178a55cbaf6-kube-api-access-mx9bx\") pod \"migrator-59844c95c7-pvf8l\" (UID: \"09809668-72b4-4a16-a821-3178a55cbaf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256085 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-979xg\" (UniqueName: \"kubernetes.io/projected/05cc0657-182f-4683-85af-eae7f03745b1-kube-api-access-979xg\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256169 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256195 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4klnh\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-kube-api-access-4klnh\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256219 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mv69\" (UniqueName: \"kubernetes.io/projected/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-kube-api-access-8mv69\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256242 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-plugins-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256265 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z67bt\" (UniqueName: \"kubernetes.io/projected/68656fd7-dcfb-4eba-b5a5-ff442f39b963-kube-api-access-z67bt\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256416 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hpdv\" (UniqueName: \"kubernetes.io/projected/963713f9-3b62-40e3-8735-16438185ac17-kube-api-access-7hpdv\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256446 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-service-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256489 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-mountpoint-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256568 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-service-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256594 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-default-certificate\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256636 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256659 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-registration-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256700 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzf8q\" (UniqueName: \"kubernetes.io/projected/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-kube-api-access-qzf8q\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256725 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/28391950-7de5-486a-8394-5a96de836f07-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256750 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256790 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cjc4\" (UniqueName: \"kubernetes.io/projected/92cea25b-5193-412c-bd80-eb4932e7a7a5-kube-api-access-4cjc4\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256847 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwpgx\" (UniqueName: \"kubernetes.io/projected/300e4ea6-63ca-4b08-9386-995aa6a67b5a-kube-api-access-qwpgx\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.256891 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257482 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257629 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68656fd7-dcfb-4eba-b5a5-ff442f39b963-metrics-tls\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257720 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-service-ca-bundle\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257758 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-srv-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257884 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/528c6e94-51ea-42a2-bb57-f521453e46bf-trusted-ca\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.257959 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-metrics-certs\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258009 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eeef0f55-17ef-4532-b2ab-a4e6feecf597-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258061 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqv62\" (UniqueName: \"kubernetes.io/projected/abca25a5-e96f-4e09-8a43-ab855b44d7a2-kube-api-access-dqv62\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258088 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxpr8\" (UniqueName: \"kubernetes.io/projected/2fb0c14a-689b-494a-80b8-ce1df08a29e5-kube-api-access-zxpr8\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258123 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qjhs\" (UniqueName: \"kubernetes.io/projected/ac99467a-7988-41f3-96bf-908859713734-kube-api-access-5qjhs\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258155 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-config\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258240 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70bd79c6-78d6-4da2-8868-40694cb22fe6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258282 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxkzq\" (UniqueName: \"kubernetes.io/projected/dc211904-b2bb-4dc9-90cb-4e30abdd981f-kube-api-access-rxkzq\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.258546 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-trusted-ca\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.259395 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.259436 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-node-bootstrap-token\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.259457 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68656fd7-dcfb-4eba-b5a5-ff442f39b963-config-volume\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260114 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac99467a-7988-41f3-96bf-908859713734-serving-cert\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260126 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac99467a-7988-41f3-96bf-908859713734-trusted-ca\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260205 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260353 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8szz\" (UniqueName: \"kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260401 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260471 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/963713f9-3b62-40e3-8735-16438185ac17-signing-key\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260501 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h8lj\" (UniqueName: \"kubernetes.io/projected/8de3abe3-bac9-4295-a7c6-b26c6e69d927-kube-api-access-6h8lj\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260572 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeef0f55-17ef-4532-b2ab-a4e6feecf597-config\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260595 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/963713f9-3b62-40e3-8735-16438185ac17-signing-cabundle\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260618 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb0c14a-689b-494a-80b8-ce1df08a29e5-config\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260651 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeef0f55-17ef-4532-b2ab-a4e6feecf597-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260723 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260932 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdcgz\" (UniqueName: \"kubernetes.io/projected/cd8a6658-4df0-49a1-8853-a038631b2bf3-kube-api-access-cdcgz\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.260988 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-client\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.261059 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.761023891 +0000 UTC m=+139.120792498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.263872 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.265960 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.273438 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-encryption-config\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.286204 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.286432 4836 projected.go:194] Error preparing data for projected volume kube-api-access-v68xd for pod openshift-controller-manager/controller-manager-879f6c89f-mfxbj: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.287495 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.787471681 +0000 UTC m=+139.147240198 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-v68xd" (UniqueName: "kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.303825 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.312575 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d81482-c475-480e-ad24-1e5eee403a99-serving-cert\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.325512 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.334797 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.345183 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.349501 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-audit\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.369911 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370095 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-socket-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370125 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70bd79c6-78d6-4da2-8868-40694cb22fe6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370146 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e4ea6-63ca-4b08-9386-995aa6a67b5a-serving-cert\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370168 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-csi-data-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370199 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370237 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd8a6658-4df0-49a1-8853-a038631b2bf3-metrics-tls\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370267 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370289 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370311 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrc6h\" (UniqueName: \"kubernetes.io/projected/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-kube-api-access-xrc6h\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370332 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370352 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-srv-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370372 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.370393 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.371030 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.372276 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.872260791 +0000 UTC m=+139.232029308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372354 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372379 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5q8r\" (UniqueName: \"kubernetes.io/projected/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-kube-api-access-x5q8r\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372421 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372436 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-socket-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372444 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx9bx\" (UniqueName: \"kubernetes.io/projected/09809668-72b4-4a16-a821-3178a55cbaf6-kube-api-access-mx9bx\") pod \"migrator-59844c95c7-pvf8l\" (UID: \"09809668-72b4-4a16-a821-3178a55cbaf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372469 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-979xg\" (UniqueName: \"kubernetes.io/projected/05cc0657-182f-4683-85af-eae7f03745b1-kube-api-access-979xg\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372490 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-stats-auth\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372519 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mv69\" (UniqueName: \"kubernetes.io/projected/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-kube-api-access-8mv69\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372537 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-plugins-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372555 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z67bt\" (UniqueName: \"kubernetes.io/projected/68656fd7-dcfb-4eba-b5a5-ff442f39b963-kube-api-access-z67bt\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372573 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hpdv\" (UniqueName: \"kubernetes.io/projected/963713f9-3b62-40e3-8735-16438185ac17-kube-api-access-7hpdv\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372588 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-service-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372606 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-mountpoint-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372621 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-service-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372635 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-default-certificate\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372662 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372677 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-registration-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372696 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzf8q\" (UniqueName: \"kubernetes.io/projected/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-kube-api-access-qzf8q\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372711 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/28391950-7de5-486a-8394-5a96de836f07-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372726 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372772 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cjc4\" (UniqueName: \"kubernetes.io/projected/92cea25b-5193-412c-bd80-eb4932e7a7a5-kube-api-access-4cjc4\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372786 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwpgx\" (UniqueName: \"kubernetes.io/projected/300e4ea6-63ca-4b08-9386-995aa6a67b5a-kube-api-access-qwpgx\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372812 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372829 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372844 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68656fd7-dcfb-4eba-b5a5-ff442f39b963-metrics-tls\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372863 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-service-ca-bundle\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372878 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-srv-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372901 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-metrics-certs\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372918 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eeef0f55-17ef-4532-b2ab-a4e6feecf597-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372934 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqv62\" (UniqueName: \"kubernetes.io/projected/abca25a5-e96f-4e09-8a43-ab855b44d7a2-kube-api-access-dqv62\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372950 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxpr8\" (UniqueName: \"kubernetes.io/projected/2fb0c14a-689b-494a-80b8-ce1df08a29e5-kube-api-access-zxpr8\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372973 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-config\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.372989 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70bd79c6-78d6-4da2-8868-40694cb22fe6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373008 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxkzq\" (UniqueName: \"kubernetes.io/projected/dc211904-b2bb-4dc9-90cb-4e30abdd981f-kube-api-access-rxkzq\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373029 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-node-bootstrap-token\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373058 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68656fd7-dcfb-4eba-b5a5-ff442f39b963-config-volume\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373090 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8szz\" (UniqueName: \"kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373106 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373124 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373139 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h8lj\" (UniqueName: \"kubernetes.io/projected/8de3abe3-bac9-4295-a7c6-b26c6e69d927-kube-api-access-6h8lj\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373156 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/963713f9-3b62-40e3-8735-16438185ac17-signing-key\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373174 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb0c14a-689b-494a-80b8-ce1df08a29e5-config\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373191 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeef0f55-17ef-4532-b2ab-a4e6feecf597-config\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373206 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/963713f9-3b62-40e3-8735-16438185ac17-signing-cabundle\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373224 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeef0f55-17ef-4532-b2ab-a4e6feecf597-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373253 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373270 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdcgz\" (UniqueName: \"kubernetes.io/projected/cd8a6658-4df0-49a1-8853-a038631b2bf3-kube-api-access-cdcgz\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373285 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-client\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373311 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stmhf\" (UniqueName: \"kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373330 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373348 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x646h\" (UniqueName: \"kubernetes.io/projected/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-kube-api-access-x646h\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373366 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92cea25b-5193-412c-bd80-eb4932e7a7a5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.373386 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28391950-7de5-486a-8394-5a96de836f07-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.378005 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.378668 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-registration-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.379948 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70bd79c6-78d6-4da2-8868-40694cb22fe6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.380068 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-csi-data-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.381174 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cd8a6658-4df0-49a1-8853-a038631b2bf3-metrics-tls\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.381223 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-plugins-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.381419 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8de3abe3-bac9-4295-a7c6-b26c6e69d927-mountpoint-dir\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.381923 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-service-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.384484 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.384747 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:15.884728824 +0000 UTC m=+139.244497341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.385714 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.387447 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/68656fd7-dcfb-4eba-b5a5-ff442f39b963-metrics-tls\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.388169 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-service-ca-bundle\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.395331 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-metrics-certs\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.397387 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-srv-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.398256 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.399729 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68656fd7-dcfb-4eba-b5a5-ff442f39b963-config-volume\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.403078 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.405650 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-config\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.409188 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-service-ca\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.409196 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.410721 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-default-certificate\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.413182 4836 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.413352 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config podName:e66435af-b690-40a3-a039-63a541178060 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.413307074 +0000 UTC m=+139.773075591 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config") pod "controller-manager-879f6c89f-mfxbj" (UID: "e66435af-b690-40a3-a039-63a541178060") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.413616 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.414871 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-apiservice-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.415470 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70bd79c6-78d6-4da2-8868-40694cb22fe6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416333 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-config\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416399 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb0c14a-689b-494a-80b8-ce1df08a29e5-serving-cert\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416433 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n89t5\" (UniqueName: \"kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416473 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416506 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-webhook-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416550 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416578 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8zkk\" (UniqueName: \"kubernetes.io/projected/70bd79c6-78d6-4da2-8868-40694cb22fe6-kube-api-access-w8zkk\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416604 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-proxy-tls\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416635 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdr8h\" (UniqueName: \"kubernetes.io/projected/6a3ebaf3-caf9-43e8-aa07-221af07bff44-kube-api-access-vdr8h\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416655 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416680 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416706 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc211904-b2bb-4dc9-90cb-4e30abdd981f-cert\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416732 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqb4v\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-kube-api-access-sqb4v\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416758 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcgtk\" (UniqueName: \"kubernetes.io/projected/368ef53c-70f8-4c26-b844-b45a07cc28eb-kube-api-access-bcgtk\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416782 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/abca25a5-e96f-4e09-8a43-ab855b44d7a2-tmpfs\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416827 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-certs\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416864 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-serving-cert\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.416859 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb0c14a-689b-494a-80b8-ce1df08a29e5-config\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.417386 4836 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.417711 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle podName:09d81482-c475-480e-ad24-1e5eee403a99 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.417684077 +0000 UTC m=+139.777452594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle") pod "apiserver-76f77b778f-txqw8" (UID: "09d81482-c475-480e-ad24-1e5eee403a99") : failed to sync configmap cache: timed out waiting for the condition Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.418431 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-profile-collector-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.418779 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.420374 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.421201 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/28391950-7de5-486a-8394-5a96de836f07-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.421757 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.421801 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-stats-auth\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.422275 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/300e4ea6-63ca-4b08-9386-995aa6a67b5a-serving-cert\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.422673 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-node-bootstrap-token\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.423082 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/963713f9-3b62-40e3-8735-16438185ac17-signing-key\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.423237 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eeef0f55-17ef-4532-b2ab-a4e6feecf597-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.423507 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb0c14a-689b-494a-80b8-ce1df08a29e5-serving-cert\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.423749 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.423861 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-profile-collector-cert\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.424006 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/abca25a5-e96f-4e09-8a43-ab855b44d7a2-tmpfs\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.424995 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.425256 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/300e4ea6-63ca-4b08-9386-995aa6a67b5a-config\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.425694 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeef0f55-17ef-4532-b2ab-a4e6feecf597-config\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.426337 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-etcd-client\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.426507 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/963713f9-3b62-40e3-8735-16438185ac17-signing-cabundle\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.426772 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.429316 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-proxy-tls\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.430422 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.433939 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.435371 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/abca25a5-e96f-4e09-8a43-ab855b44d7a2-webhook-cert\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.435549 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.435808 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/05cc0657-182f-4683-85af-eae7f03745b1-certs\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.436201 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/368ef53c-70f8-4c26-b844-b45a07cc28eb-srv-cert\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.438166 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc211904-b2bb-4dc9-90cb-4e30abdd981f-cert\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.439326 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a3ebaf3-caf9-43e8-aa07-221af07bff44-serving-cert\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.440350 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.440976 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/92cea25b-5193-412c-bd80-eb4932e7a7a5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.447266 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28391950-7de5-486a-8394-5a96de836f07-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.491970 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ggf\" (UniqueName: \"kubernetes.io/projected/c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45-kube-api-access-g6ggf\") pod \"downloads-7954f5f757-z9w5n\" (UID: \"c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45\") " pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.503664 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.518480 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.519894 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.019865251 +0000 UTC m=+139.379633768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.525900 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krspr\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.529941 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8wj\" (UniqueName: \"kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.530067 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.530568 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.030554513 +0000 UTC m=+139.390323030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.535915 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8wj\" (UniqueName: \"kubernetes.io/projected/09d81482-c475-480e-ad24-1e5eee403a99-kube-api-access-xg8wj\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.545361 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.549620 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.561369 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4klnh\" (UniqueName: \"kubernetes.io/projected/528c6e94-51ea-42a2-bb57-f521453e46bf-kube-api-access-4klnh\") pod \"ingress-operator-5b745b69d9-hjr8w\" (UID: \"528c6e94-51ea-42a2-bb57-f521453e46bf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.582307 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qjhs\" (UniqueName: \"kubernetes.io/projected/ac99467a-7988-41f3-96bf-908859713734-kube-api-access-5qjhs\") pod \"console-operator-58897d9998-b9fsf\" (UID: \"ac99467a-7988-41f3-96bf-908859713734\") " pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.602031 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx9bx\" (UniqueName: \"kubernetes.io/projected/09809668-72b4-4a16-a821-3178a55cbaf6-kube-api-access-mx9bx\") pod \"migrator-59844c95c7-pvf8l\" (UID: \"09809668-72b4-4a16-a821-3178a55cbaf6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.626626 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwpgx\" (UniqueName: \"kubernetes.io/projected/300e4ea6-63ca-4b08-9386-995aa6a67b5a-kube-api-access-qwpgx\") pod \"authentication-operator-69f744f599-pw59n\" (UID: \"300e4ea6-63ca-4b08-9386-995aa6a67b5a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.630698 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.630924 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.130892274 +0000 UTC m=+139.490660801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.631501 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.631905 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.631968 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.131957215 +0000 UTC m=+139.491725742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.645836 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-979xg\" (UniqueName: \"kubernetes.io/projected/05cc0657-182f-4683-85af-eae7f03745b1-kube-api-access-979xg\") pod \"machine-config-server-lll5j\" (UID: \"05cc0657-182f-4683-85af-eae7f03745b1\") " pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.647911 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.687247 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hpdv\" (UniqueName: \"kubernetes.io/projected/963713f9-3b62-40e3-8735-16438185ac17-kube-api-access-7hpdv\") pod \"service-ca-9c57cc56f-zlfc7\" (UID: \"963713f9-3b62-40e3-8735-16438185ac17\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.690548 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z67bt\" (UniqueName: \"kubernetes.io/projected/68656fd7-dcfb-4eba-b5a5-ff442f39b963-kube-api-access-z67bt\") pod \"dns-default-c7hkn\" (UID: \"68656fd7-dcfb-4eba-b5a5-ff442f39b963\") " pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.705918 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8490d8e-8273-4b7a-94cd-85b7aa91f79a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fqcjv\" (UID: \"a8490d8e-8273-4b7a-94cd-85b7aa91f79a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.706293 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.721295 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.728963 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.733157 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.736547 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.236528106 +0000 UTC m=+139.596296613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.740626 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqv62\" (UniqueName: \"kubernetes.io/projected/abca25a5-e96f-4e09-8a43-ab855b44d7a2-kube-api-access-dqv62\") pod \"packageserver-d55dfcdfc-g6wfj\" (UID: \"abca25a5-e96f-4e09-8a43-ab855b44d7a2\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.747189 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.749164 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h8lj\" (UniqueName: \"kubernetes.io/projected/8de3abe3-bac9-4295-a7c6-b26c6e69d927-kube-api-access-6h8lj\") pod \"csi-hostpathplugin-s7pq7\" (UID: \"8de3abe3-bac9-4295-a7c6-b26c6e69d927\") " pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.761887 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.772011 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mv69\" (UniqueName: \"kubernetes.io/projected/5cd2e5f4-c2ab-46c0-966d-bdb43059ec00-kube-api-access-8mv69\") pod \"olm-operator-6b444d44fb-fkzkm\" (UID: \"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.780169 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.784773 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eeef0f55-17ef-4532-b2ab-a4e6feecf597-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7krzd\" (UID: \"eeef0f55-17ef-4532-b2ab-a4e6feecf597\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: W0122 09:44:15.784848 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac27fcac_0bc1_41a1_aac0_2fdce1d06c04.slice/crio-ce399616d6a18a4bb4d5374dcb888e21f304395a7762439fee17c696300c60e3 WatchSource:0}: Error finding container ce399616d6a18a4bb4d5374dcb888e21f304395a7762439fee17c696300c60e3: Status 404 returned error can't find the container with id ce399616d6a18a4bb4d5374dcb888e21f304395a7762439fee17c696300c60e3 Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.786953 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.795131 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.801524 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdcgz\" (UniqueName: \"kubernetes.io/projected/cd8a6658-4df0-49a1-8853-a038631b2bf3-kube-api-access-cdcgz\") pod \"dns-operator-744455d44c-88xxc\" (UID: \"cd8a6658-4df0-49a1-8853-a038631b2bf3\") " pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.823835 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stmhf\" (UniqueName: \"kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf\") pod \"collect-profiles-29484570-hjs8r\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.831104 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.834898 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.835014 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.835434 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.335418476 +0000 UTC m=+139.695187043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.839428 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.842027 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.843158 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cjc4\" (UniqueName: \"kubernetes.io/projected/92cea25b-5193-412c-bd80-eb4932e7a7a5-kube-api-access-4cjc4\") pod \"control-plane-machine-set-operator-78cbb6b69f-stt4s\" (UID: \"92cea25b-5193-412c-bd80-eb4932e7a7a5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.846050 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.847526 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" event={"ID":"4cbc633b-ae56-438d-88c3-91437b545f38","Type":"ContainerStarted","Data":"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.848271 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.848963 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tbm5x" event={"ID":"37f85569-398f-4d97-8971-958d0010e4d4","Type":"ContainerStarted","Data":"2b019d42321cb4461643884fa37721896518a0ae7fce4a6c4b27d630488e30d9"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.849812 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" event={"ID":"a588b581-1bb9-4c23-8be2-26816172ddeb","Type":"ContainerStarted","Data":"2bd5ebae46df9d9c23aed4256af3bbf8ffafbbf99d707427b6d8e7579c9f8d70"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.851098 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" event={"ID":"ca83435d-4c56-403e-90a9-3afd1fbd417d","Type":"ContainerStarted","Data":"de2d669ecf7f4daf51bbee9897601701b3c0ef829dab6b2e128e6ff75f12f8ca"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.858342 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" event={"ID":"dc87ab7f-7d2d-48da-bb1a-a5905f8cb40a","Type":"ContainerStarted","Data":"8dad382526652cae2e9441e88c68df62f025ce5412a626ffd457447208095bbc"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.864497 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxkzq\" (UniqueName: \"kubernetes.io/projected/dc211904-b2bb-4dc9-90cb-4e30abdd981f-kube-api-access-rxkzq\") pod \"ingress-canary-hdgd6\" (UID: \"dc211904-b2bb-4dc9-90cb-4e30abdd981f\") " pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.867289 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.874815 4836 generic.go:334] "Generic (PLEG): container finished" podID="172b6aec-4183-447f-82d0-5904cbbc7604" containerID="fbdc2dcd32cac74e6de9dc2abf2d0dcc300fc9ceffb1ab3d0508f5dd4434fd14" exitCode=0 Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.875604 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" event={"ID":"172b6aec-4183-447f-82d0-5904cbbc7604","Type":"ContainerDied","Data":"fbdc2dcd32cac74e6de9dc2abf2d0dcc300fc9ceffb1ab3d0508f5dd4434fd14"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.887209 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.892870 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" event={"ID":"42863163-25f6-41d7-953b-804ba3e052c8","Type":"ContainerStarted","Data":"b18cedbf0fd1e526a448b98e537dc8077b2e9628b4bae4ed13c8cd9d1e8d57f9"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.892909 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.892925 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxpr8\" (UniqueName: \"kubernetes.io/projected/2fb0c14a-689b-494a-80b8-ce1df08a29e5-kube-api-access-zxpr8\") pod \"service-ca-operator-777779d784-p9xhq\" (UID: \"2fb0c14a-689b-494a-80b8-ce1df08a29e5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.899404 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" event={"ID":"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec","Type":"ContainerStarted","Data":"c626d61f96275795f01cda6337300f1f70bbb1cefccd19f48ef9d3591ed843df"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.899427 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" event={"ID":"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec","Type":"ContainerStarted","Data":"cda5be83e31d103fa2e7bc12aa56261c1fd6ca2cef30bb639f866786045a837c"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.900931 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" event={"ID":"1121b7db-f2c3-4900-81c7-e765a19a091c","Type":"ContainerStarted","Data":"58d4cb24dba2e2f6036262392271bb97e8b199da2b9d07bc2cb3e80484f4c65e"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.900981 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" event={"ID":"1121b7db-f2c3-4900-81c7-e765a19a091c","Type":"ContainerStarted","Data":"84a108d49d1d771d33e266bebb6f1fccb6dfd12fecddd4a7b8f1a87cfab89c15"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.900996 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" event={"ID":"1121b7db-f2c3-4900-81c7-e765a19a091c","Type":"ContainerStarted","Data":"ac90b651ffc3e1561971cbe5f43d737b74a56d977ee9b7b430597dc7eece2c57"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.903346 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" event={"ID":"f517dfdc-296c-41c3-8506-dc95aa8504db","Type":"ContainerStarted","Data":"ac2a661a04293d732943fd8aa8ce78514eb043e3f2a4161c257ecee1e3966cbd"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.904746 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" event={"ID":"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04","Type":"ContainerStarted","Data":"ce399616d6a18a4bb4d5374dcb888e21f304395a7762439fee17c696300c60e3"} Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.905444 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8szz\" (UniqueName: \"kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz\") pod \"route-controller-manager-6576b87f9c-m66s2\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.917365 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.917660 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrc6h\" (UniqueName: \"kubernetes.io/projected/4fe19b48-498b-420a-9a84-9b63a0a1ae9e-kube-api-access-xrc6h\") pod \"multus-admission-controller-857f4d67dd-g5pts\" (UID: \"4fe19b48-498b-420a-9a84-9b63a0a1ae9e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.922502 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.929475 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lll5j" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.935337 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hdgd6" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.938289 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzf8q\" (UniqueName: \"kubernetes.io/projected/b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4-kube-api-access-qzf8q\") pod \"machine-config-controller-84d6567774-c84j7\" (UID: \"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.938518 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.938654 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.438637538 +0000 UTC m=+139.798406055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.938962 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:15 crc kubenswrapper[4836]: E0122 09:44:15.942564 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.4425474 +0000 UTC m=+139.802316017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.944929 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-z9w5n"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.958911 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x646h\" (UniqueName: \"kubernetes.io/projected/5afa5ef0-f947-4bf6-82b2-99b47e25e10d-kube-api-access-x646h\") pod \"router-default-5444994796-9mq8w\" (UID: \"5afa5ef0-f947-4bf6-82b2-99b47e25e10d\") " pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.984485 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5q8r\" (UniqueName: \"kubernetes.io/projected/5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f-kube-api-access-x5q8r\") pod \"package-server-manager-789f6589d5-n6njz\" (UID: \"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.984509 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.986962 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b9fsf"] Jan 22 09:44:15 crc kubenswrapper[4836]: I0122 09:44:15.998065 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.016983 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n89t5\" (UniqueName: \"kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5\") pod \"marketplace-operator-79b997595-4bdcx\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.033371 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.037030 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcgtk\" (UniqueName: \"kubernetes.io/projected/368ef53c-70f8-4c26-b844-b45a07cc28eb-kube-api-access-bcgtk\") pod \"catalog-operator-68c6474976-fqmdw\" (UID: \"368ef53c-70f8-4c26-b844-b45a07cc28eb\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.040102 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.040224 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.540208944 +0000 UTC m=+139.899977461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.040513 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.040829 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.540819982 +0000 UTC m=+139.900588509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.056129 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqb4v\" (UniqueName: \"kubernetes.io/projected/28391950-7de5-486a-8394-5a96de836f07-kube-api-access-sqb4v\") pod \"cluster-image-registry-operator-dc59b4c8b-vvgf2\" (UID: \"28391950-7de5-486a-8394-5a96de836f07\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.058720 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.098110 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8zkk\" (UniqueName: \"kubernetes.io/projected/70bd79c6-78d6-4da2-8868-40694cb22fe6-kube-api-access-w8zkk\") pod \"openshift-apiserver-operator-796bbdcf4f-5r86k\" (UID: \"70bd79c6-78d6-4da2-8868-40694cb22fe6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.101834 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.109522 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.117737 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.124933 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.142112 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.142259 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.642229514 +0000 UTC m=+140.001998051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.142467 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.143017 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.643001935 +0000 UTC m=+140.002770452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.153491 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.162349 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.174853 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.183707 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.244300 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.244637 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.744622592 +0000 UTC m=+140.104391109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.340321 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.345478 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.345740 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.845727785 +0000 UTC m=+140.205496302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.447126 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.447328 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.447410 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.947369254 +0000 UTC m=+140.307137811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.447581 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.448506 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09d81482-c475-480e-ad24-1e5eee403a99-trusted-ca-bundle\") pod \"apiserver-76f77b778f-txqw8\" (UID: \"09d81482-c475-480e-ad24-1e5eee403a99\") " pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.448790 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"controller-manager-879f6c89f-mfxbj\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.549588 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.549919 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.049908487 +0000 UTC m=+140.409677004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.594208 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdr8h\" (UniqueName: \"kubernetes.io/projected/6a3ebaf3-caf9-43e8-aa07-221af07bff44-kube-api-access-vdr8h\") pod \"etcd-operator-b45778765-wp857\" (UID: \"6a3ebaf3-caf9-43e8-aa07-221af07bff44\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.628543 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" Jan 22 09:44:16 crc kubenswrapper[4836]: W0122 09:44:16.632276 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod528c6e94_51ea_42a2_bb57_f521453e46bf.slice/crio-729ae54fbdd59e6272cd8378a330c43ea1129af5ddc5e0fe675a6221bb6fc559 WatchSource:0}: Error finding container 729ae54fbdd59e6272cd8378a330c43ea1129af5ddc5e0fe675a6221bb6fc559: Status 404 returned error can't find the container with id 729ae54fbdd59e6272cd8378a330c43ea1129af5ddc5e0fe675a6221bb6fc559 Jan 22 09:44:16 crc kubenswrapper[4836]: W0122 09:44:16.650632 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac99467a_7988_41f3_96bf_908859713734.slice/crio-944f0ad7b394f43907632257bee3bf9c804c13bd593d581a6336641b4d9fd78d WatchSource:0}: Error finding container 944f0ad7b394f43907632257bee3bf9c804c13bd593d581a6336641b4d9fd78d: Status 404 returned error can't find the container with id 944f0ad7b394f43907632257bee3bf9c804c13bd593d581a6336641b4d9fd78d Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.650851 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.651278 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.151258728 +0000 UTC m=+140.511027255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.662476 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.679226 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.714369 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.753199 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.753600 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.253585685 +0000 UTC m=+140.613354222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.858051 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.859190 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-btclw" podStartSLOduration=120.859171124 podStartE2EDuration="2m0.859171124s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:16.8568977 +0000 UTC m=+140.216666227" watchObservedRunningTime="2026-01-22 09:44:16.859171124 +0000 UTC m=+140.218939641" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.860116 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.36009243 +0000 UTC m=+140.719860947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.862088 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv"] Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.913946 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" event={"ID":"de54019b-ce0c-4dd3-8c94-121cf6d0d5ec","Type":"ContainerStarted","Data":"062606c6e44724ef5fe2e5b0c78b048a139e931fbb75b7eeadaed413d2866f9b"} Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.914777 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" event={"ID":"528c6e94-51ea-42a2-bb57-f521453e46bf","Type":"ContainerStarted","Data":"729ae54fbdd59e6272cd8378a330c43ea1129af5ddc5e0fe675a6221bb6fc559"} Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.915591 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" event={"ID":"ac99467a-7988-41f3-96bf-908859713734","Type":"ContainerStarted","Data":"944f0ad7b394f43907632257bee3bf9c804c13bd593d581a6336641b4d9fd78d"} Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.917330 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z9w5n" event={"ID":"c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45","Type":"ContainerStarted","Data":"596a39faee65479e1cca6727b7c964e64cf0995b35a5cfcb8a7d276f8a25dc46"} Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.961213 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:16 crc kubenswrapper[4836]: E0122 09:44:16.961534 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.461516652 +0000 UTC m=+140.821285189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:16 crc kubenswrapper[4836]: I0122 09:44:16.978248 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pw59n"] Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.002124 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g5pts"] Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.036349 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.061958 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.062246 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.562215233 +0000 UTC m=+140.921983750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.063744 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.064264 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.564246271 +0000 UTC m=+140.924014798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.138117 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s"] Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.164801 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.165502 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.665482867 +0000 UTC m=+141.025251404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: W0122 09:44:17.178196 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod300e4ea6_63ca_4b08_9386_995aa6a67b5a.slice/crio-ede7b2cd0e56e9bedac6e45d460bfcc97f38c7fe27e41fb11e78a4717dc6d54d WatchSource:0}: Error finding container ede7b2cd0e56e9bedac6e45d460bfcc97f38c7fe27e41fb11e78a4717dc6d54d: Status 404 returned error can't find the container with id ede7b2cd0e56e9bedac6e45d460bfcc97f38c7fe27e41fb11e78a4717dc6d54d Jan 22 09:44:17 crc kubenswrapper[4836]: W0122 09:44:17.181213 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fe19b48_498b_420a_9a84_9b63a0a1ae9e.slice/crio-7246e5a415f24c60029e029cd69914853a91cf5b5b5bace0e03c26982bfcee29 WatchSource:0}: Error finding container 7246e5a415f24c60029e029cd69914853a91cf5b5b5bace0e03c26982bfcee29: Status 404 returned error can't find the container with id 7246e5a415f24c60029e029cd69914853a91cf5b5b5bace0e03c26982bfcee29 Jan 22 09:44:17 crc kubenswrapper[4836]: W0122 09:44:17.225420 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05cc0657_182f_4683_85af_eae7f03745b1.slice/crio-b88b35017d1acfddb06803452815d4ad45761ca4c092fde6d376f20bab2cc078 WatchSource:0}: Error finding container b88b35017d1acfddb06803452815d4ad45761ca4c092fde6d376f20bab2cc078: Status 404 returned error can't find the container with id b88b35017d1acfddb06803452815d4ad45761ca4c092fde6d376f20bab2cc078 Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.267099 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.267386 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.767375662 +0000 UTC m=+141.127144179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.368162 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.368487 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.868456195 +0000 UTC m=+141.228224722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.368897 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.369255 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.869239267 +0000 UTC m=+141.229007784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.382812 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" podStartSLOduration=122.382792461 podStartE2EDuration="2m2.382792461s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:17.381685099 +0000 UTC m=+140.741453616" watchObservedRunningTime="2026-01-22 09:44:17.382792461 +0000 UTC m=+140.742560978" Jan 22 09:44:17 crc kubenswrapper[4836]: W0122 09:44:17.436357 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5afa5ef0_f947_4bf6_82b2_99b47e25e10d.slice/crio-c9a27649e0215cf1e810bd3c2ed3ccce3712245174cee19953c6abc6daea5599 WatchSource:0}: Error finding container c9a27649e0215cf1e810bd3c2ed3ccce3712245174cee19953c6abc6daea5599: Status 404 returned error can't find the container with id c9a27649e0215cf1e810bd3c2ed3ccce3712245174cee19953c6abc6daea5599 Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.471147 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.471649 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:17.971629986 +0000 UTC m=+141.331398503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.572762 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.573516 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.073499951 +0000 UTC m=+141.433268468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.673955 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.674264 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.174249884 +0000 UTC m=+141.534018391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.699303 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6jmp5" podStartSLOduration=121.699277023 podStartE2EDuration="2m1.699277023s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:17.696135984 +0000 UTC m=+141.055904501" watchObservedRunningTime="2026-01-22 09:44:17.699277023 +0000 UTC m=+141.059045540" Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.774989 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.775390 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.275377487 +0000 UTC m=+141.635146004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.876637 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.876847 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.37681999 +0000 UTC m=+141.736588507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.877542 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.878047 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.377986722 +0000 UTC m=+141.737755229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.925827 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lll5j" event={"ID":"05cc0657-182f-4683-85af-eae7f03745b1","Type":"ContainerStarted","Data":"b88b35017d1acfddb06803452815d4ad45761ca4c092fde6d376f20bab2cc078"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.928723 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" event={"ID":"ac27fcac-0bc1-41a1-aac0-2fdce1d06c04","Type":"ContainerStarted","Data":"5fb7cbc89a14306e2a80847cefe048e76047b9dfd7bd6c6af94d2ddd5a39a8a6"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.934336 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" event={"ID":"528c6e94-51ea-42a2-bb57-f521453e46bf","Type":"ContainerStarted","Data":"f45cf609050d6ec99e56b2d0b1e91c386b08e5438dd9f3246541d6656838b946"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.937552 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" event={"ID":"e66435af-b690-40a3-a039-63a541178060","Type":"ContainerStarted","Data":"d15eebeacd158a8e66f082cecda80d5930ccabc5e827b892c7fd0c28183a744a"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.938490 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" event={"ID":"4fe19b48-498b-420a-9a84-9b63a0a1ae9e","Type":"ContainerStarted","Data":"7246e5a415f24c60029e029cd69914853a91cf5b5b5bace0e03c26982bfcee29"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.942383 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" event={"ID":"42863163-25f6-41d7-953b-804ba3e052c8","Type":"ContainerStarted","Data":"a3f7526f0a23e928d46db229442465fc4ec476fd544ff9cbbc93d4143dff5dd7"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.944654 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" event={"ID":"ca83435d-4c56-403e-90a9-3afd1fbd417d","Type":"ContainerStarted","Data":"2572752c7236fb6edfa6f5bddb9b5b8271cc1e90127966ddfd356bb739570783"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.946535 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" event={"ID":"a588b581-1bb9-4c23-8be2-26816172ddeb","Type":"ContainerStarted","Data":"51431a2ae30234bfcb308895f742404603709075dddf5d26c1e4a4f622c0bac3"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.958078 4836 generic.go:334] "Generic (PLEG): container finished" podID="f517dfdc-296c-41c3-8506-dc95aa8504db" containerID="b65e48c2feb80d1c875876c94cc283dfc022575c4a272f6b34906ab7116e0be9" exitCode=0 Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.958178 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" event={"ID":"f517dfdc-296c-41c3-8506-dc95aa8504db","Type":"ContainerDied","Data":"b65e48c2feb80d1c875876c94cc283dfc022575c4a272f6b34906ab7116e0be9"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.961615 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" event={"ID":"a8490d8e-8273-4b7a-94cd-85b7aa91f79a","Type":"ContainerStarted","Data":"d0a3b160feccb8dfd256bebfb92435f0118a284bc3735d219ae9cdf47c2807df"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.962749 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9mq8w" event={"ID":"5afa5ef0-f947-4bf6-82b2-99b47e25e10d","Type":"ContainerStarted","Data":"c9a27649e0215cf1e810bd3c2ed3ccce3712245174cee19953c6abc6daea5599"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.964014 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" event={"ID":"92cea25b-5193-412c-bd80-eb4932e7a7a5","Type":"ContainerStarted","Data":"65e33ebfb06a3bc2c21926e2a83d2f3208c28c8f92b30d78b6ea38a6385ef48d"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.965242 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" event={"ID":"ac99467a-7988-41f3-96bf-908859713734","Type":"ContainerStarted","Data":"5114707a974b73dbd848bd9fcc1b04be3f93c4cebe1dbe189a0f97b0678e4ceb"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.966903 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tbm5x" event={"ID":"37f85569-398f-4d97-8971-958d0010e4d4","Type":"ContainerStarted","Data":"4a5b7cccb4349646461681a49b4fef0afaad02d65742313b74187899f0f63635"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.968142 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" event={"ID":"300e4ea6-63ca-4b08-9386-995aa6a67b5a","Type":"ContainerStarted","Data":"ede7b2cd0e56e9bedac6e45d460bfcc97f38c7fe27e41fb11e78a4717dc6d54d"} Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.978481 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.978793 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.478773526 +0000 UTC m=+141.838542043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:17 crc kubenswrapper[4836]: I0122 09:44:17.978920 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:17 crc kubenswrapper[4836]: E0122 09:44:17.979366 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.479308252 +0000 UTC m=+141.839076769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.084225 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.084930 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.584907292 +0000 UTC m=+141.944675859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.088878 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.089163 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.589150142 +0000 UTC m=+141.948918659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.099119 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-fkgzv" podStartSLOduration=123.099099343 podStartE2EDuration="2m3.099099343s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:18.058452803 +0000 UTC m=+141.418221340" watchObservedRunningTime="2026-01-22 09:44:18.099099343 +0000 UTC m=+141.458867850" Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.130282 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-frpzv" podStartSLOduration=122.130264897 podStartE2EDuration="2m2.130264897s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:18.079839739 +0000 UTC m=+141.439608276" watchObservedRunningTime="2026-01-22 09:44:18.130264897 +0000 UTC m=+141.490033414" Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.130379 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pn4sd" podStartSLOduration=123.13037496 podStartE2EDuration="2m3.13037496s" podCreationTimestamp="2026-01-22 09:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:18.12087199 +0000 UTC m=+141.480640507" watchObservedRunningTime="2026-01-22 09:44:18.13037496 +0000 UTC m=+141.490143477" Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.173199 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-tbm5x" podStartSLOduration=122.173182692 podStartE2EDuration="2m2.173182692s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:18.172433481 +0000 UTC m=+141.532201998" watchObservedRunningTime="2026-01-22 09:44:18.173182692 +0000 UTC m=+141.532951209" Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.196965 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.197469 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.697450498 +0000 UTC m=+142.057219005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.247190 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qvxjj" podStartSLOduration=122.247175506 podStartE2EDuration="2m2.247175506s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:18.245058356 +0000 UTC m=+141.604826883" watchObservedRunningTime="2026-01-22 09:44:18.247175506 +0000 UTC m=+141.606944013" Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.292942 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s7pq7"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.299767 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.300122 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.800108785 +0000 UTC m=+142.159877302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.401147 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.401348 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.901322811 +0000 UTC m=+142.261091328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.401604 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.401932 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:18.901919658 +0000 UTC m=+142.261688175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.502604 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.502745 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.002719822 +0000 UTC m=+142.362488339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.502890 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.503323 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.00331565 +0000 UTC m=+142.363084167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.580276 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c7hkn"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.604767 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.604896 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.104875715 +0000 UTC m=+142.464644232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.605335 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.605725 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.105708339 +0000 UTC m=+142.465476856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.706806 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.706963 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.206941595 +0000 UTC m=+142.566710112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.707108 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.707413 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.207401859 +0000 UTC m=+142.567170376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.727911 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.731716 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.733060 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.735214 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.752330 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw"] Jan 22 09:44:18 crc kubenswrapper[4836]: W0122 09:44:18.757404 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf699f3_dacd_4f97_97b3_0bb300377623.slice/crio-71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380 WatchSource:0}: Error finding container 71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380: Status 404 returned error can't find the container with id 71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380 Jan 22 09:44:18 crc kubenswrapper[4836]: W0122 09:44:18.766860 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb39f4e2d_e53e_4a33_9c8b_9775c6c6c4f4.slice/crio-762001b2a1c7a4530aef6e6ef830fa89e27bfef6818b6ff636bcde0d291bc41c WatchSource:0}: Error finding container 762001b2a1c7a4530aef6e6ef830fa89e27bfef6818b6ff636bcde0d291bc41c: Status 404 returned error can't find the container with id 762001b2a1c7a4530aef6e6ef830fa89e27bfef6818b6ff636bcde0d291bc41c Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.783125 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.783218 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.787104 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-txqw8"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.793062 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hdgd6"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.794027 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlfc7"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.814409 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.814765 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.314750638 +0000 UTC m=+142.674519155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.917557 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:18 crc kubenswrapper[4836]: E0122 09:44:18.918028 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.418013242 +0000 UTC m=+142.777781759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.932306 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-88xxc"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.965124 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.965169 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.969070 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2"] Jan 22 09:44:18 crc kubenswrapper[4836]: I0122 09:44:18.975258 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wp857"] Jan 22 09:44:19 crc kubenswrapper[4836]: W0122 09:44:19.012958 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a3ebaf3_caf9_43e8_aa07_221af07bff44.slice/crio-90a650e4233d3acbd8aa1b018a65187ef91fb979b3816864e5ef70744669349f WatchSource:0}: Error finding container 90a650e4233d3acbd8aa1b018a65187ef91fb979b3816864e5ef70744669349f: Status 404 returned error can't find the container with id 90a650e4233d3acbd8aa1b018a65187ef91fb979b3816864e5ef70744669349f Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.019442 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.019540 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.019801 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.519787164 +0000 UTC m=+142.879555681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.021161 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.049742 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" event={"ID":"f517dfdc-296c-41c3-8506-dc95aa8504db","Type":"ContainerStarted","Data":"30ba0e80a086ef015f87664efa97a93ca0c10a4eb4ea025c5860a22d1e877e38"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.066855 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.068960 4836 csr.go:261] certificate signing request csr-qlk98 is approved, waiting to be issued Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.068975 4836 csr.go:257] certificate signing request csr-qlk98 is issued Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.074925 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k"] Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.101031 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" podStartSLOduration=123.101013584 podStartE2EDuration="2m3.101013584s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.096870497 +0000 UTC m=+142.456639014" watchObservedRunningTime="2026-01-22 09:44:19.101013584 +0000 UTC m=+142.460782101" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.128193 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" event={"ID":"2fb0c14a-689b-494a-80b8-ce1df08a29e5","Type":"ContainerStarted","Data":"5fd8322c473f128fdb1637fea13efff7344c392a3c5307106ffbddde0f420c7b"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.148221 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" event={"ID":"ca83435d-4c56-403e-90a9-3afd1fbd417d","Type":"ContainerStarted","Data":"42cc771c63ea2a8316cb3a40f23f1c6998ef52234311b7063c00fd1ecbc88e2b"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.164997 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.171619 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.671594292 +0000 UTC m=+143.031362809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.183812 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-g5kkn" podStartSLOduration=123.183792728 podStartE2EDuration="2m3.183792728s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.181444402 +0000 UTC m=+142.541212909" watchObservedRunningTime="2026-01-22 09:44:19.183792728 +0000 UTC m=+142.543561245" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.194569 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" event={"ID":"8de3abe3-bac9-4295-a7c6-b26c6e69d927","Type":"ContainerStarted","Data":"5f09bf8cc8168d1676324f971541093395da0239bdabd3e9e154bb6298a79af0"} Jan 22 09:44:19 crc kubenswrapper[4836]: W0122 09:44:19.194784 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9368ed_7945_471a_9869_aa2fdb8ad2c7.slice/crio-d19eefba4f8fce67a7636a84a2b27458f3e34dc75e2855e05caad53d0af07f9f WatchSource:0}: Error finding container d19eefba4f8fce67a7636a84a2b27458f3e34dc75e2855e05caad53d0af07f9f: Status 404 returned error can't find the container with id d19eefba4f8fce67a7636a84a2b27458f3e34dc75e2855e05caad53d0af07f9f Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.228405 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" event={"ID":"eeef0f55-17ef-4532-b2ab-a4e6feecf597","Type":"ContainerStarted","Data":"4e3320501a8022527a276ae1963cec65f726cc5fc58d4d3720b72a7ff50c8579"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.239114 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" event={"ID":"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00","Type":"ContainerStarted","Data":"5bfb1661a2962dd6695c437a01bfe8a18f1eaf9f27fbffd860c563eaefe81995"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.251771 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" event={"ID":"92cea25b-5193-412c-bd80-eb4932e7a7a5","Type":"ContainerStarted","Data":"0c379ba01e1488ddab568668f191dd3205ffc00642e3c1ceac1d05a3abe70e94"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.261894 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" event={"ID":"faf699f3-dacd-4f97-97b3-0bb300377623","Type":"ContainerStarted","Data":"71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.268561 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.269373 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.769351681 +0000 UTC m=+143.129120198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.277962 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-stt4s" podStartSLOduration=123.277948734 podStartE2EDuration="2m3.277948734s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.275674199 +0000 UTC m=+142.635442716" watchObservedRunningTime="2026-01-22 09:44:19.277948734 +0000 UTC m=+142.637717251" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.301056 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" event={"ID":"e66435af-b690-40a3-a039-63a541178060","Type":"ContainerStarted","Data":"4455a6b2e0b04552492f4b39c68ccb873bd23797ca9c94f3a9542e36e40c6fbe"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.301668 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.305937 4836 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mfxbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.305991 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.307939 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c7hkn" event={"ID":"68656fd7-dcfb-4eba-b5a5-ff442f39b963","Type":"ContainerStarted","Data":"8a7f94699caad55ebb6bae0a12c0a937873341a1596ab6ca7c0864678bad09a4"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.318864 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podStartSLOduration=123.318842852 podStartE2EDuration="2m3.318842852s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.317315279 +0000 UTC m=+142.677083796" watchObservedRunningTime="2026-01-22 09:44:19.318842852 +0000 UTC m=+142.678611369" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.327824 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" event={"ID":"a8490d8e-8273-4b7a-94cd-85b7aa91f79a","Type":"ContainerStarted","Data":"5fc914d175176f0529762d046bd2e08883573b99b778f2082c519df99a76e3e6"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.333154 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" event={"ID":"300e4ea6-63ca-4b08-9386-995aa6a67b5a","Type":"ContainerStarted","Data":"0bc26178dad517fbd5c4f6c6631ff83093c482acc4cde53bdb21330105a6d8d7"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.339157 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" event={"ID":"528c6e94-51ea-42a2-bb57-f521453e46bf","Type":"ContainerStarted","Data":"28a0bc772b765bd50d1d3365b91d28ebdc931267bc14e1b0178e6df41d771fdb"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.343956 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fqcjv" podStartSLOduration=123.343945103 podStartE2EDuration="2m3.343945103s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.34348436 +0000 UTC m=+142.703252887" watchObservedRunningTime="2026-01-22 09:44:19.343945103 +0000 UTC m=+142.703713620" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.344492 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" event={"ID":"4fe19b48-498b-420a-9a84-9b63a0a1ae9e","Type":"ContainerStarted","Data":"27b828d61f97d871611647dfee57ac1cc39420bf3e176ea56cb000f919a4ebd9"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.345656 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-z9w5n" event={"ID":"c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45","Type":"ContainerStarted","Data":"efb35d0340b4e4ed77ddfb026c29bd8778b3eb8f1c9233f27c307ca40f67bb3c"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.345791 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.346375 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" event={"ID":"963713f9-3b62-40e3-8735-16438185ac17","Type":"ContainerStarted","Data":"97928f1f961fbe3aaeca12c62f58944191f83905d98352c9567b309514d2fa0f"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.346940 4836 patch_prober.go:28] interesting pod/downloads-7954f5f757-z9w5n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.346974 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z9w5n" podUID="c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.349401 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9mq8w" event={"ID":"5afa5ef0-f947-4bf6-82b2-99b47e25e10d","Type":"ContainerStarted","Data":"309dfa6730e8a07bbf57fd1e4736d30f76c1595fe19af7053a4ef7aaf8264777"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.350670 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lll5j" event={"ID":"05cc0657-182f-4683-85af-eae7f03745b1","Type":"ContainerStarted","Data":"3c7a049b47ea93a5cc2a795571b06f26a686c4260ea95ddaa1cccd75b5449e52"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.351355 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" event={"ID":"09d81482-c475-480e-ad24-1e5eee403a99","Type":"ContainerStarted","Data":"d4b62f21a0bc94f20aa15e0799be6406ce93ac244b846e98e76c064d61256b86"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.352144 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" event={"ID":"abca25a5-e96f-4e09-8a43-ab855b44d7a2","Type":"ContainerStarted","Data":"bf353efe5b73711abd411e29e52f6f2eeb0b06752cce0124635cf53efb55eb01"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.355371 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" event={"ID":"172b6aec-4183-447f-82d0-5904cbbc7604","Type":"ContainerStarted","Data":"d7ebcfac20f9a3a713d05adac40a1f6a778e9cf087c9b7991a7e5b6e00bdfe17"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.364790 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hjr8w" podStartSLOduration=123.364780053 podStartE2EDuration="2m3.364780053s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.364544406 +0000 UTC m=+142.724312923" watchObservedRunningTime="2026-01-22 09:44:19.364780053 +0000 UTC m=+142.724548570" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.366436 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" event={"ID":"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4","Type":"ContainerStarted","Data":"762001b2a1c7a4530aef6e6ef830fa89e27bfef6818b6ff636bcde0d291bc41c"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.369941 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.371167 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.871154613 +0000 UTC m=+143.230923120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.389100 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" event={"ID":"368ef53c-70f8-4c26-b844-b45a07cc28eb","Type":"ContainerStarted","Data":"6ca6148518f6cb3ed5a4ca2566a7a880cd36b2145ff7953d2f4bf2e46bb79487"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.393509 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hdgd6" event={"ID":"dc211904-b2bb-4dc9-90cb-4e30abdd981f","Type":"ContainerStarted","Data":"1a842594d8fd34ddc6beb4118a79c7e67e020bef4e40b25387aeb66166418ed3"} Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.398030 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.401152 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pw59n" podStartSLOduration=123.401126632 podStartE2EDuration="2m3.401126632s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.392426966 +0000 UTC m=+142.752195483" watchObservedRunningTime="2026-01-22 09:44:19.401126632 +0000 UTC m=+142.760895149" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.401985 4836 patch_prober.go:28] interesting pod/console-operator-58897d9998-b9fsf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.402020 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" podUID="ac99467a-7988-41f3-96bf-908859713734" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/readyz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.413866 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" podStartSLOduration=122.413851353 podStartE2EDuration="2m2.413851353s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.412933746 +0000 UTC m=+142.772702263" watchObservedRunningTime="2026-01-22 09:44:19.413851353 +0000 UTC m=+142.773619860" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.426065 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-z9w5n" podStartSLOduration=123.426032727 podStartE2EDuration="2m3.426032727s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.425303517 +0000 UTC m=+142.785072034" watchObservedRunningTime="2026-01-22 09:44:19.426032727 +0000 UTC m=+142.785801244" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.456718 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9mq8w" podStartSLOduration=123.456691425 podStartE2EDuration="2m3.456691425s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.453879256 +0000 UTC m=+142.813647773" watchObservedRunningTime="2026-01-22 09:44:19.456691425 +0000 UTC m=+142.816459942" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.465877 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lll5j" podStartSLOduration=7.465860795 podStartE2EDuration="7.465860795s" podCreationTimestamp="2026-01-22 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.464930099 +0000 UTC m=+142.824698606" watchObservedRunningTime="2026-01-22 09:44:19.465860795 +0000 UTC m=+142.825629312" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.471601 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.473010 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:19.972992417 +0000 UTC m=+143.332760934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.481789 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" podStartSLOduration=123.481768405 podStartE2EDuration="2m3.481768405s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:19.479991425 +0000 UTC m=+142.839759962" watchObservedRunningTime="2026-01-22 09:44:19.481768405 +0000 UTC m=+142.841536922" Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.573291 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.573622 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.073594556 +0000 UTC m=+143.433363063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.674365 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.674553 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.174526344 +0000 UTC m=+143.534294871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.675031 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.675412 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.175395048 +0000 UTC m=+143.535163565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.776483 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.776884 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.276862961 +0000 UTC m=+143.636631488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.878441 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.878902 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.37888258 +0000 UTC m=+143.738651127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.980794 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.981162 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.481133555 +0000 UTC m=+143.840902082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:19 crc kubenswrapper[4836]: I0122 09:44:19.981332 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:19 crc kubenswrapper[4836]: E0122 09:44:19.981950 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.481938778 +0000 UTC m=+143.841707315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.059317 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.064666 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:20 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:20 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:20 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.064721 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.073116 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-22 09:39:19 +0000 UTC, rotation deadline is 2026-12-10 09:13:03.35368675 +0000 UTC Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.073152 4836 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7727h28m43.280537409s for next certificate rotation Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.083518 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.083788 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.583768872 +0000 UTC m=+143.943537389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.186766 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.187181 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.68716166 +0000 UTC m=+144.046930177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.288305 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.288441 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.788421627 +0000 UTC m=+144.148190144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.288757 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.289238 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.789218179 +0000 UTC m=+144.148986766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.399227 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.400099 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:20.900072148 +0000 UTC m=+144.259840685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.418102 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" event={"ID":"abca25a5-e96f-4e09-8a43-ab855b44d7a2","Type":"ContainerStarted","Data":"86a2eebb99b5a895a6d0b33ce75fe8a3e3760c00eb74a2c263aed18aa4eebb52"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.422471 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" event={"ID":"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f","Type":"ContainerStarted","Data":"b2003311dce33cfa40d827a01daec83aebe308ebe0a7fa84c90b12a35b1cddb3"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.423539 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" event={"ID":"28391950-7de5-486a-8394-5a96de836f07","Type":"ContainerStarted","Data":"34396f9fb2ca8a748b287baa7785b8d8142e7800f7bc9b79d5c2e7ef32ea2cee"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.424783 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" event={"ID":"5148a0f2-8687-4dd0-af2c-58cd45924d13","Type":"ContainerStarted","Data":"bba508c3fb7b38ccfd628a9586f625afe5c348f822d5013cb17c52550be57161"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.425833 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" event={"ID":"6a3ebaf3-caf9-43e8-aa07-221af07bff44","Type":"ContainerStarted","Data":"90a650e4233d3acbd8aa1b018a65187ef91fb979b3816864e5ef70744669349f"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.426890 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" event={"ID":"09809668-72b4-4a16-a821-3178a55cbaf6","Type":"ContainerStarted","Data":"0bfe067e8f06d4fd93417652d36820fe0df71838e9e5694c4ce376fbf4c32030"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.427569 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerStarted","Data":"d19eefba4f8fce67a7636a84a2b27458f3e34dc75e2855e05caad53d0af07f9f"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.430187 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" event={"ID":"cd8a6658-4df0-49a1-8853-a038631b2bf3","Type":"ContainerStarted","Data":"4eb7f20fea48a754989ceee91aec57bb34f93f9ccc05655f0d76ddd863bf7af4"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.432841 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" event={"ID":"70bd79c6-78d6-4da2-8868-40694cb22fe6","Type":"ContainerStarted","Data":"ae03af9b7e6205f6e5611b3af4c65e13444431dea3c959543dfeb22093bb6128"} Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.437058 4836 patch_prober.go:28] interesting pod/downloads-7954f5f757-z9w5n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.437126 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z9w5n" podUID="c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.437461 4836 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mfxbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.437524 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.500810 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.501194 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.001178261 +0000 UTC m=+144.360946778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.601540 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.601766 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.101736529 +0000 UTC m=+144.461505056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.603594 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.604026 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.104009513 +0000 UTC m=+144.463778120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.705909 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.706187 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.206150875 +0000 UTC m=+144.565919432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.706320 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.706773 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.206757922 +0000 UTC m=+144.566526479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.807420 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.807700 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.307669599 +0000 UTC m=+144.667438156 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.807788 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.808125 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.308112672 +0000 UTC m=+144.667881189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.908397 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.908583 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.408556326 +0000 UTC m=+144.768324853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:20 crc kubenswrapper[4836]: I0122 09:44:20.908748 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:20 crc kubenswrapper[4836]: E0122 09:44:20.909354 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.409339639 +0000 UTC m=+144.769108166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.009708 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.010098 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.510081252 +0000 UTC m=+144.869849769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.062407 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:21 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:21 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:21 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.062466 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.068411 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vd5ft" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.111405 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.111687 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.611673948 +0000 UTC m=+144.971442465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.212746 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.213178 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.713159991 +0000 UTC m=+145.072928508 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.313744 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.314111 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.814091929 +0000 UTC m=+145.173860516 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.387689 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-b9fsf" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.416730 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.417133 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:21.917099746 +0000 UTC m=+145.276868273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.451168 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" event={"ID":"5cd2e5f4-c2ab-46c0-966d-bdb43059ec00","Type":"ContainerStarted","Data":"efefa4767ac248bebc0344f3c818b87fdf83ed480fd32e511e2a16d074893b4e"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.451365 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.454135 4836 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-fkzkm container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.454187 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" podUID="5cd2e5f4-c2ab-46c0-966d-bdb43059ec00" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.470379 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" event={"ID":"faf699f3-dacd-4f97-97b3-0bb300377623","Type":"ContainerStarted","Data":"8587bb5ed5bba69f8434b325c4f4f0ef7ed60a0a00673650469a642ab5a88328"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.517329 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.518090 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.018071495 +0000 UTC m=+145.377840012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.529181 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" event={"ID":"368ef53c-70f8-4c26-b844-b45a07cc28eb","Type":"ContainerStarted","Data":"efda6eb24ac21adc410d8c52160397a6b59833e80e55a5abb6e0a9642d7558ad"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.529981 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.556287 4836 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fqmdw container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.556338 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" podUID="368ef53c-70f8-4c26-b844-b45a07cc28eb" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.557418 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" podStartSLOduration=125.557406299 podStartE2EDuration="2m5.557406299s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.501145276 +0000 UTC m=+144.860913803" watchObservedRunningTime="2026-01-22 09:44:21.557406299 +0000 UTC m=+144.917174816" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.558745 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" podStartSLOduration=125.558739176 podStartE2EDuration="2m5.558739176s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.558601943 +0000 UTC m=+144.918370460" watchObservedRunningTime="2026-01-22 09:44:21.558739176 +0000 UTC m=+144.918507713" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.572169 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" event={"ID":"eeef0f55-17ef-4532-b2ab-a4e6feecf597","Type":"ContainerStarted","Data":"90f0f415928a7b05f0f93624755589d5761f1dcc39ea4acda2464e135b957261"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.573840 4836 generic.go:334] "Generic (PLEG): container finished" podID="09d81482-c475-480e-ad24-1e5eee403a99" containerID="50dda9d49960c64354b7f31492bb64fc178fb646a0eb464578d2f946cf47ab40" exitCode=0 Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.573894 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" event={"ID":"09d81482-c475-480e-ad24-1e5eee403a99","Type":"ContainerDied","Data":"50dda9d49960c64354b7f31492bb64fc178fb646a0eb464578d2f946cf47ab40"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.574837 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" event={"ID":"5148a0f2-8687-4dd0-af2c-58cd45924d13","Type":"ContainerStarted","Data":"63b6238ad7a62b56c78627f4021f1b06eefec5ec90a01d055f76f76de6bfd38c"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.575571 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.584375 4836 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m66s2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.584422 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.586650 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" podStartSLOduration=125.586639997 podStartE2EDuration="2m5.586639997s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.584585499 +0000 UTC m=+144.944354016" watchObservedRunningTime="2026-01-22 09:44:21.586639997 +0000 UTC m=+144.946408514" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.618600 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.618784 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.118754136 +0000 UTC m=+145.478522653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.618923 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.620993 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.12098043 +0000 UTC m=+145.480748947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.625259 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podStartSLOduration=124.62523703 podStartE2EDuration="2m4.62523703s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.624245292 +0000 UTC m=+144.984013809" watchObservedRunningTime="2026-01-22 09:44:21.62523703 +0000 UTC m=+144.985005547" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.641838 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" event={"ID":"963713f9-3b62-40e3-8735-16438185ac17","Type":"ContainerStarted","Data":"9c2a566aa7ef1c150597b4fae397eb842418cecbd7a04677067f33e2fc875695"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.655927 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7krzd" podStartSLOduration=125.655907188 podStartE2EDuration="2m5.655907188s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.655701452 +0000 UTC m=+145.015469979" watchObservedRunningTime="2026-01-22 09:44:21.655907188 +0000 UTC m=+145.015675715" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.668085 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" event={"ID":"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4","Type":"ContainerStarted","Data":"48f5618c343f54ae1fef864b3ab5cd1bd5556415343b31ba6999a42a9d953b76"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.698011 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" event={"ID":"70bd79c6-78d6-4da2-8868-40694cb22fe6","Type":"ContainerStarted","Data":"3c149da2f1f4279256939244d3749483aac10d3432dc7990131380deb6ffdf3f"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.720635 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.722160 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.222124154 +0000 UTC m=+145.581892671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.728942 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zlfc7" podStartSLOduration=124.723493342 podStartE2EDuration="2m4.723493342s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.698896025 +0000 UTC m=+145.058664552" watchObservedRunningTime="2026-01-22 09:44:21.723493342 +0000 UTC m=+145.083261859" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.731592 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5r86k" podStartSLOduration=125.731570331 podStartE2EDuration="2m5.731570331s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.729711508 +0000 UTC m=+145.089480025" watchObservedRunningTime="2026-01-22 09:44:21.731570331 +0000 UTC m=+145.091338848" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.733530 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c7hkn" event={"ID":"68656fd7-dcfb-4eba-b5a5-ff442f39b963","Type":"ContainerStarted","Data":"fdf9512c291d84f7e154e7c09375052a8d3c0563dc1e9affa798985519f34a66"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.743762 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" event={"ID":"2fb0c14a-689b-494a-80b8-ce1df08a29e5","Type":"ContainerStarted","Data":"a5d73696b74ed40371afaadd40756a8dadadbe2c1d82a4ccd1c87fff793ce941"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.756938 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" event={"ID":"cd8a6658-4df0-49a1-8853-a038631b2bf3","Type":"ContainerStarted","Data":"077f228ee20a3790f182fd848502aa43e6466c0dd26fb5d3c2f1efd24b64dfa0"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.767667 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" event={"ID":"28391950-7de5-486a-8394-5a96de836f07","Type":"ContainerStarted","Data":"db4b288256db392846a0d3dcbc21e0cd98668ac34b3b03c40f073b97564d5d0c"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.775177 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" event={"ID":"09809668-72b4-4a16-a821-3178a55cbaf6","Type":"ContainerStarted","Data":"0b4fcf8636ada97ff66fb532d97138c6fb475e58991b17f5be128b30196cac9d"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.803815 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-vvgf2" podStartSLOduration=125.803785266 podStartE2EDuration="2m5.803785266s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.80358967 +0000 UTC m=+145.163358187" watchObservedRunningTime="2026-01-22 09:44:21.803785266 +0000 UTC m=+145.163553783" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.804082 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p9xhq" podStartSLOduration=124.804075084 podStartE2EDuration="2m4.804075084s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.781411692 +0000 UTC m=+145.141180219" watchObservedRunningTime="2026-01-22 09:44:21.804075084 +0000 UTC m=+145.163843621" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.813699 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" event={"ID":"4fe19b48-498b-420a-9a84-9b63a0a1ae9e","Type":"ContainerStarted","Data":"5aacc285b17341cda4cee9e4d0167c978774c0305222a4b3650c4d3b8ebfa6f5"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.827106 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.827429 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.327413864 +0000 UTC m=+145.687182381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.837306 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerStarted","Data":"54eb9b908456f56ece981ae792ccb16c3e8150b01e7d353f2f834607d8d287e6"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.839107 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.843714 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" event={"ID":"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f","Type":"ContainerStarted","Data":"b49b4a03caa05efe6edb789682e5aa877a0541e00707b94b1c0e505f810d1513"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.857199 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4bdcx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.857255 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.859381 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-g5pts" podStartSLOduration=125.85936556 podStartE2EDuration="2m5.85936556s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.857524247 +0000 UTC m=+145.217292764" watchObservedRunningTime="2026-01-22 09:44:21.85936556 +0000 UTC m=+145.219134077" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.902623 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" event={"ID":"6a3ebaf3-caf9-43e8-aa07-221af07bff44","Type":"ContainerStarted","Data":"96ccd03b460f0c6046aafcc290c72f48e141a51eb59269fca17cafa33fa8d971"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.928506 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.929051 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.429015572 +0000 UTC m=+145.788784089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.931857 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hdgd6" event={"ID":"dc211904-b2bb-4dc9-90cb-4e30abdd981f","Type":"ContainerStarted","Data":"907d7e8aeae919f6af4b29c012962f180c345208bb4ee377e44ab24b45fdfb41"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.932468 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:21 crc kubenswrapper[4836]: E0122 09:44:21.932778 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.432766468 +0000 UTC m=+145.792534985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.942691 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podStartSLOduration=124.942678438 podStartE2EDuration="2m4.942678438s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.892504867 +0000 UTC m=+145.252273374" watchObservedRunningTime="2026-01-22 09:44:21.942678438 +0000 UTC m=+145.302446955" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.943972 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-wp857" podStartSLOduration=125.943967365 podStartE2EDuration="2m5.943967365s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.942888544 +0000 UTC m=+145.302657071" watchObservedRunningTime="2026-01-22 09:44:21.943967365 +0000 UTC m=+145.303735882" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.948987 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" event={"ID":"8de3abe3-bac9-4295-a7c6-b26c6e69d927","Type":"ContainerStarted","Data":"ae8993752b7af7a6fe26e6d57a6a737a4166ad817bc750fd11bb902126e04f0c"} Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.949030 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.970501 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hdgd6" podStartSLOduration=9.970483985 podStartE2EDuration="9.970483985s" podCreationTimestamp="2026-01-22 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.967345487 +0000 UTC m=+145.327114004" watchObservedRunningTime="2026-01-22 09:44:21.970483985 +0000 UTC m=+145.330252502" Jan 22 09:44:21 crc kubenswrapper[4836]: I0122 09:44:21.981930 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.027263 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" podStartSLOduration=125.027244833 podStartE2EDuration="2m5.027244833s" podCreationTimestamp="2026-01-22 09:42:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:21.994453675 +0000 UTC m=+145.354222192" watchObservedRunningTime="2026-01-22 09:44:22.027244833 +0000 UTC m=+145.387013350" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.033214 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.034312 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.534297523 +0000 UTC m=+145.894066040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.065389 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:22 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:22 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:22 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.065647 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.134659 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.135187 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.635171599 +0000 UTC m=+145.994940106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.236377 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.236724 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.736709844 +0000 UTC m=+146.096478361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.337695 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.338113 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.838092475 +0000 UTC m=+146.197861042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.363154 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g6wfj" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.439450 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.439638 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.93960873 +0000 UTC m=+146.299377247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.439777 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.440133 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:22.940118694 +0000 UTC m=+146.299887211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.541070 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.541390 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.041374771 +0000 UTC m=+146.401143288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.642954 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.643306 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.143291087 +0000 UTC m=+146.503059604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.744426 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.744595 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.244561125 +0000 UTC m=+146.604329662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.744746 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.745031 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.245018298 +0000 UTC m=+146.604786815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.846611 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.847115 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.347098338 +0000 UTC m=+146.706866855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.948471 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:22 crc kubenswrapper[4836]: E0122 09:44:22.948810 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.448799078 +0000 UTC m=+146.808567595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.953646 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" event={"ID":"cd8a6658-4df0-49a1-8853-a038631b2bf3","Type":"ContainerStarted","Data":"7719b41a092a052cd73532b8bc038b8832754c27fac649f233900bf5e1e9968c"} Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.956687 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" event={"ID":"b39f4e2d-e53e-4a33-9c8b-9775c6c6c4f4","Type":"ContainerStarted","Data":"fa1d98313a11577f95988153eb98aa1264a0d24f498f33209979ff031823d8e6"} Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.958738 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" event={"ID":"09809668-72b4-4a16-a821-3178a55cbaf6","Type":"ContainerStarted","Data":"bf2aad0ad2a510a04b0403f8878ffdc178e7ba6fdf3b5793f12b0942accfba61"} Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.960247 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c7hkn" event={"ID":"68656fd7-dcfb-4eba-b5a5-ff442f39b963","Type":"ContainerStarted","Data":"3737bf04945089dc885bfee869f9ac9c08df0d3c913f6299c100d205c2f5aceb"} Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.961686 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" event={"ID":"5d630b5c-cdee-4bc4-9d9c-80acc9fc7c6f","Type":"ContainerStarted","Data":"90343515f514781b2a95fefbd190a43426b17075b7ca253ca3652058551ecb6d"} Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.962476 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4bdcx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.962517 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.964016 4836 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m66s2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.964103 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.979394 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-fkzkm" Jan 22 09:44:22 crc kubenswrapper[4836]: I0122 09:44:22.992854 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-88xxc" podStartSLOduration=126.992836424 podStartE2EDuration="2m6.992836424s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:22.991909808 +0000 UTC m=+146.351678345" watchObservedRunningTime="2026-01-22 09:44:22.992836424 +0000 UTC m=+146.352604951" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.027690 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c84j7" podStartSLOduration=127.027670431 podStartE2EDuration="2m7.027670431s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:23.027188698 +0000 UTC m=+146.386957215" watchObservedRunningTime="2026-01-22 09:44:23.027670431 +0000 UTC m=+146.387438948" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.049987 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.051257 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.551236878 +0000 UTC m=+146.911005395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.068445 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:23 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:23 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:23 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.068494 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.110996 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pvf8l" podStartSLOduration=127.110977209 podStartE2EDuration="2m7.110977209s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:23.107162322 +0000 UTC m=+146.466930849" watchObservedRunningTime="2026-01-22 09:44:23.110977209 +0000 UTC m=+146.470745726" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.152274 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.154720 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.654704798 +0000 UTC m=+147.014473415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.155821 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" podStartSLOduration=127.155807119 podStartE2EDuration="2m7.155807119s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:23.13818298 +0000 UTC m=+146.497951527" watchObservedRunningTime="2026-01-22 09:44:23.155807119 +0000 UTC m=+146.515575636" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.253921 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.254122 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.754092533 +0000 UTC m=+147.113861050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.254438 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.254943 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.754932896 +0000 UTC m=+147.114701523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.355829 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.356279 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.856260285 +0000 UTC m=+147.216028812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.417712 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fqmdw" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.457981 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.458308 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:23.958296954 +0000 UTC m=+147.318065471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.559004 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.559388 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.059371677 +0000 UTC m=+147.419140194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.660333 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.660629 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.160617824 +0000 UTC m=+147.520386341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.761436 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.761608 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.261583693 +0000 UTC m=+147.621352210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.761838 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.762153 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.262141898 +0000 UTC m=+147.621910405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.840304 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.840551 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.849993 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.862767 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.863222 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.36320191 +0000 UTC m=+147.722970437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.897418 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.897475 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.964166 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:23 crc kubenswrapper[4836]: E0122 09:44:23.965009 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.464993212 +0000 UTC m=+147.824761739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.977980 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" event={"ID":"8de3abe3-bac9-4295-a7c6-b26c6e69d927","Type":"ContainerStarted","Data":"55d9e55d51a94e5b6858000780f6402a5f7c6577c810d5bf2903546b430a4f4d"} Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.978981 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.979537 4836 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m66s2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.979573 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.979677 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4bdcx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.979723 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Jan 22 09:44:23 crc kubenswrapper[4836]: I0122 09:44:23.991170 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-lsxgt" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.008027 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-c7hkn" podStartSLOduration=12.008008441 podStartE2EDuration="12.008008441s" podCreationTimestamp="2026-01-22 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:24.005666104 +0000 UTC m=+147.365434621" watchObservedRunningTime="2026-01-22 09:44:24.008008441 +0000 UTC m=+147.367776958" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.061905 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:24 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:24 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:24 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.061965 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.065482 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.065627 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.565608741 +0000 UTC m=+147.925377258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.065728 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.065777 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.065822 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.065864 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.066215 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.566199408 +0000 UTC m=+147.925967925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.066298 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.071115 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.071123 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.083067 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.104519 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.167068 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.167224 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.667197197 +0000 UTC m=+148.026965714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.167684 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.168827 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.668809523 +0000 UTC m=+148.028578120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.213297 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.222016 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.230076 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.268543 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.268696 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.768670031 +0000 UTC m=+148.128438548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.268741 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.269209 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.769199716 +0000 UTC m=+148.128968233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.370542 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.371357 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.871338168 +0000 UTC m=+148.231106695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.473997 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.474344 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:24.974333405 +0000 UTC m=+148.334101922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.575132 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.575380 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.075354425 +0000 UTC m=+148.435122942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.575470 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.575760 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.075743826 +0000 UTC m=+148.435512343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.676434 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.676598 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.176576501 +0000 UTC m=+148.536345018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.676673 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.676959 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.176950961 +0000 UTC m=+148.536719478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.777917 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.778125 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.278098856 +0000 UTC m=+148.637867373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.778233 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.778524 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.278512248 +0000 UTC m=+148.638280765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.878959 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.879157 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.379133837 +0000 UTC m=+148.738902354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.879250 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.879592 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.379570019 +0000 UTC m=+148.739338536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.923504 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:24 crc kubenswrapper[4836]: W0122 09:44:24.938193 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-7369f4d19319cc7d2af0e763b4bf0e7eccfbb44fb46c4a094e0887309505a872 WatchSource:0}: Error finding container 7369f4d19319cc7d2af0e763b4bf0e7eccfbb44fb46c4a094e0887309505a872: Status 404 returned error can't find the container with id 7369f4d19319cc7d2af0e763b4bf0e7eccfbb44fb46c4a094e0887309505a872 Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.980533 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.980659 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.480640851 +0000 UTC m=+148.840409368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.981104 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:24 crc kubenswrapper[4836]: E0122 09:44:24.981430 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.481419223 +0000 UTC m=+148.841187750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.990919 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"10c32ce297dba401ddda50ee6e8717264852a33eefabf671601a97fb85b611db"} Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.992637 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"646d9a60f99b5164985dda1b3f792ebbab35f293af94f47065eb4f67bc3a86a3"} Jan 22 09:44:24 crc kubenswrapper[4836]: I0122 09:44:24.993922 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7369f4d19319cc7d2af0e763b4bf0e7eccfbb44fb46c4a094e0887309505a872"} Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.063941 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:25 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:25 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:25 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.063999 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.080027 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.080090 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.081760 4836 patch_prober.go:28] interesting pod/console-f9d7485db-tbm5x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.081825 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-tbm5x" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.081970 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.082183 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.582160316 +0000 UTC m=+148.941928833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.083534 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.083802 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.583793702 +0000 UTC m=+148.943562219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.184963 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.185172 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.685132452 +0000 UTC m=+149.044900969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.186305 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.186682 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.686659784 +0000 UTC m=+149.046428351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.287302 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.287525 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.78749068 +0000 UTC m=+149.147259207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.287660 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.287998 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.787982914 +0000 UTC m=+149.147751431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.388906 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.389106 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.889085977 +0000 UTC m=+149.248854504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.389190 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.389488 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.889467408 +0000 UTC m=+149.249235925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.490069 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.490276 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.990244441 +0000 UTC m=+149.350012968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.490519 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.490877 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:25.990860379 +0000 UTC m=+149.350628896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.532610 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.533346 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.534996 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.535694 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.540579 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.592267 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.592461 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.092433634 +0000 UTC m=+149.452202151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.592820 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.593207 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.093191527 +0000 UTC m=+149.452960054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.661321 4836 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.694123 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.694322 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.694378 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.694546 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.194526936 +0000 UTC m=+149.554295453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.707936 4836 patch_prober.go:28] interesting pod/downloads-7954f5f757-z9w5n container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.707976 4836 patch_prober.go:28] interesting pod/downloads-7954f5f757-z9w5n container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.707999 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-z9w5n" podUID="c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.708016 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-z9w5n" podUID="c97cf283-b3d7-4a2c-b7d0-5b5bb7f69c45" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.789183 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.790917 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.795349 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.795751 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.795836 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.795875 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.795955 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.797122 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.29710968 +0000 UTC m=+149.656878197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.814336 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.842433 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.849321 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.896454 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.896725 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfr6h\" (UniqueName: \"kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.896753 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.896800 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:25 crc kubenswrapper[4836]: E0122 09:44:25.896898 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.396883615 +0000 UTC m=+149.756652132 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.985484 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.986402 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.992686 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:25 crc kubenswrapper[4836]: I0122 09:44:25.993953 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:25.998557 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfr6h\" (UniqueName: \"kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:25.998589 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:25.998623 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:25.998652 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:25.999674 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: E0122 09:44:26.000359 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.500348315 +0000 UTC m=+149.860116832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.002249 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.004991 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.020479 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfr6h\" (UniqueName: \"kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h\") pod \"community-operators-7bfkp\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.041713 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"892ede3f85792caecbc326b9fdb5f6956a651dd98290ee5e52216120ceb4c2e6"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.050530 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"eb33aef1e762caf1d94b969d9a4c1e8009ebd3d2b712afb8bab94fe93a489de3"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.050647 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.052747 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" event={"ID":"8de3abe3-bac9-4295-a7c6-b26c6e69d927","Type":"ContainerStarted","Data":"ec357be96488807917969a122578d4df869c4b1fceb47b686006a5f06e996b9a"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.054326 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b526d9ff6c3b38b919b2ebbd440026162c40dcc866505643f5169e9013d2d6dc"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.057827 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" event={"ID":"09d81482-c475-480e-ad24-1e5eee403a99","Type":"ContainerStarted","Data":"7de8da6d08967ae85f64fc8d5ec4b898f3a978927e95a16bb70b158b8169c5d6"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.060350 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.062479 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" event={"ID":"faf699f3-dacd-4f97-97b3-0bb300377623","Type":"ContainerDied","Data":"8587bb5ed5bba69f8434b325c4f4f0ef7ed60a0a00673650469a642ab5a88328"} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.062543 4836 generic.go:334] "Generic (PLEG): container finished" podID="faf699f3-dacd-4f97-97b3-0bb300377623" containerID="8587bb5ed5bba69f8434b325c4f4f0ef7ed60a0a00673650469a642ab5a88328" exitCode=0 Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.063831 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:26 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:26 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:26 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.063875 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.099982 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:26 crc kubenswrapper[4836]: E0122 09:44:26.100157 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.60013315 +0000 UTC m=+149.959901667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.100262 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.100345 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.100368 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.100401 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crrh\" (UniqueName: \"kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: E0122 09:44:26.101367 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:44:26.601350065 +0000 UTC m=+149.961118582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t4zmg" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.120435 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.165945 4836 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-22T09:44:25.661558282Z","Handler":null,"Name":""} Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.169390 4836 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.169422 4836 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.178202 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.179136 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.184249 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.190826 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.203086 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.203325 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crrh\" (UniqueName: \"kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.203413 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.203558 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.203971 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.205599 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.221517 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.271671 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crrh\" (UniqueName: \"kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh\") pod \"certified-operators-6pq7n\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.304940 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbnjd\" (UniqueName: \"kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.305197 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.305313 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.305377 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.305470 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.314110 4836 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.314152 4836 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.314427 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 22 09:44:26 crc kubenswrapper[4836]: W0122 09:44:26.324033 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod986b7f3a_87ee_473f_926b_104b896b7d31.slice/crio-b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4 WatchSource:0}: Error finding container b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4: Status 404 returned error can't find the container with id b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4 Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.334942 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t4zmg\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.344266 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:44:26 crc kubenswrapper[4836]: W0122 09:44:26.383014 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65b61041_36cc_4c4d_be61_dd4e370b7ff3.slice/crio-2eb611cec8de73c43e2cafcd79b7b9256db285b8f3fa1e288977f3002f349767 WatchSource:0}: Error finding container 2eb611cec8de73c43e2cafcd79b7b9256db285b8f3fa1e288977f3002f349767: Status 404 returned error can't find the container with id 2eb611cec8de73c43e2cafcd79b7b9256db285b8f3fa1e288977f3002f349767 Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.393864 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.395340 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.406784 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbnjd\" (UniqueName: \"kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.406874 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.406987 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.407854 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.414825 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.415198 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.454339 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbnjd\" (UniqueName: \"kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd\") pod \"community-operators-zbvh5\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.499526 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.508718 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvznw\" (UniqueName: \"kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.509080 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.509193 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.573354 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.612797 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvznw\" (UniqueName: \"kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.612903 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.612941 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.615489 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.616906 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.617067 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.670606 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvznw\" (UniqueName: \"kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw\") pod \"certified-operators-n4dwm\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:26 crc kubenswrapper[4836]: I0122 09:44:26.758680 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.019779 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.040390 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:44:27 crc kubenswrapper[4836]: W0122 09:44:27.047460 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b2b98a3_9481_442f_a35a_d8adcf779887.slice/crio-1eab1bac3d9313405c4a355ebc4f696142dce4b82041457bd2e5fb9ae03c1142 WatchSource:0}: Error finding container 1eab1bac3d9313405c4a355ebc4f696142dce4b82041457bd2e5fb9ae03c1142: Status 404 returned error can't find the container with id 1eab1bac3d9313405c4a355ebc4f696142dce4b82041457bd2e5fb9ae03c1142 Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.064270 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:27 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:27 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:27 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.064954 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.068442 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerStarted","Data":"282575cb2ba5fcd5ed74d9f5b2755e7eb0b92cfb5fe2253e3bc58f88ba366c61"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.075503 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" event={"ID":"8de3abe3-bac9-4295-a7c6-b26c6e69d927","Type":"ContainerStarted","Data":"f6cfd6c1b2e3068841ac7eecdbdfc6288e5f8c0c0065abe30abca7e843d05d69"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.086915 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" event={"ID":"09d81482-c475-480e-ad24-1e5eee403a99","Type":"ContainerStarted","Data":"c23a5eaaa4f93c2476c24f28819feb6ff3e16edf25f2e3eb621f1eb953106d63"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.093572 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"986b7f3a-87ee-473f-926b-104b896b7d31","Type":"ContainerStarted","Data":"b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.095112 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerStarted","Data":"1eab1bac3d9313405c4a355ebc4f696142dce4b82041457bd2e5fb9ae03c1142"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.096473 4836 generic.go:334] "Generic (PLEG): container finished" podID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerID="7a1fb1404f277b34aceb3fd88bd162f157327df6be57dcde84a7a2504c241b10" exitCode=0 Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.096606 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerDied","Data":"7a1fb1404f277b34aceb3fd88bd162f157327df6be57dcde84a7a2504c241b10"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.096635 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerStarted","Data":"2eb611cec8de73c43e2cafcd79b7b9256db285b8f3fa1e288977f3002f349767"} Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.127660 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.132896 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-s7pq7" podStartSLOduration=15.132876433 podStartE2EDuration="15.132876433s" podCreationTimestamp="2026-01-22 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:27.129330853 +0000 UTC m=+150.489099380" watchObservedRunningTime="2026-01-22 09:44:27.132876433 +0000 UTC m=+150.492644950" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.151432 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.153635 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" podStartSLOduration=131.153613691 podStartE2EDuration="2m11.153613691s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:27.152735316 +0000 UTC m=+150.512503833" watchObservedRunningTime="2026-01-22 09:44:27.153613691 +0000 UTC m=+150.513382208" Jan 22 09:44:27 crc kubenswrapper[4836]: W0122 09:44:27.369634 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cbcf6f5_ed55_4302_bdf6_ccf3c21a43c8.slice/crio-503f1b58afd615dc10a0198db457c9a8307f810f69e8c8d41f7bab0de10be715 WatchSource:0}: Error finding container 503f1b58afd615dc10a0198db457c9a8307f810f69e8c8d41f7bab0de10be715: Status 404 returned error can't find the container with id 503f1b58afd615dc10a0198db457c9a8307f810f69e8c8d41f7bab0de10be715 Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.394536 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.531608 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume\") pod \"faf699f3-dacd-4f97-97b3-0bb300377623\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.531867 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume\") pod \"faf699f3-dacd-4f97-97b3-0bb300377623\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.531888 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stmhf\" (UniqueName: \"kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf\") pod \"faf699f3-dacd-4f97-97b3-0bb300377623\" (UID: \"faf699f3-dacd-4f97-97b3-0bb300377623\") " Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.533290 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume" (OuterVolumeSpecName: "config-volume") pod "faf699f3-dacd-4f97-97b3-0bb300377623" (UID: "faf699f3-dacd-4f97-97b3-0bb300377623"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.538945 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf" (OuterVolumeSpecName: "kube-api-access-stmhf") pod "faf699f3-dacd-4f97-97b3-0bb300377623" (UID: "faf699f3-dacd-4f97-97b3-0bb300377623"). InnerVolumeSpecName "kube-api-access-stmhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.539857 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "faf699f3-dacd-4f97-97b3-0bb300377623" (UID: "faf699f3-dacd-4f97-97b3-0bb300377623"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.633677 4836 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/faf699f3-dacd-4f97-97b3-0bb300377623-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.633714 4836 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/faf699f3-dacd-4f97-97b3-0bb300377623-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.633728 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stmhf\" (UniqueName: \"kubernetes.io/projected/faf699f3-dacd-4f97-97b3-0bb300377623-kube-api-access-stmhf\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.785737 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:44:27 crc kubenswrapper[4836]: E0122 09:44:27.785941 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faf699f3-dacd-4f97-97b3-0bb300377623" containerName="collect-profiles" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.785953 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="faf699f3-dacd-4f97-97b3-0bb300377623" containerName="collect-profiles" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.786059 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="faf699f3-dacd-4f97-97b3-0bb300377623" containerName="collect-profiles" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.786720 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.788973 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.800430 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.936621 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.936688 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:27 crc kubenswrapper[4836]: I0122 09:44:27.936744 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9m69\" (UniqueName: \"kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.037947 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.038011 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.038089 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9m69\" (UniqueName: \"kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.038904 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.039146 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.046227 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.046831 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.049093 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.049108 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.059649 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9m69\" (UniqueName: \"kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69\") pod \"redhat-marketplace-ztdcc\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.061599 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:28 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:28 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:28 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.061648 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.069493 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.098738 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.106111 4836 generic.go:334] "Generic (PLEG): container finished" podID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerID="e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106" exitCode=0 Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.106198 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerDied","Data":"e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.106226 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerStarted","Data":"503f1b58afd615dc10a0198db457c9a8307f810f69e8c8d41f7bab0de10be715"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.107531 4836 generic.go:334] "Generic (PLEG): container finished" podID="986b7f3a-87ee-473f-926b-104b896b7d31" containerID="463b16ccdc4ad1ea3e944f6bf068faacf2f182da8b344a504689d82b632bcc15" exitCode=0 Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.107635 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"986b7f3a-87ee-473f-926b-104b896b7d31","Type":"ContainerDied","Data":"463b16ccdc4ad1ea3e944f6bf068faacf2f182da8b344a504689d82b632bcc15"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.109535 4836 generic.go:334] "Generic (PLEG): container finished" podID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerID="2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673" exitCode=0 Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.109606 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerDied","Data":"2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.111518 4836 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.114834 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" event={"ID":"faf699f3-dacd-4f97-97b3-0bb300377623","Type":"ContainerDied","Data":"71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.114866 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71521ad3694f9cafeff7c31d74a94dd6240107a9f7d699ca31cdfd016885a380" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.114944 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.126638 4836 generic.go:334] "Generic (PLEG): container finished" podID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerID="6d3166cd0b03f93eef234b51531da6330dcaac837a67db0bb4acd8e6762a9a6d" exitCode=0 Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.126696 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerDied","Data":"6d3166cd0b03f93eef234b51531da6330dcaac837a67db0bb4acd8e6762a9a6d"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.139196 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.139271 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.147123 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" event={"ID":"f60d64f9-1517-4093-9a99-abd333d12a36","Type":"ContainerStarted","Data":"f519e9ae5d8ad9daa7a64f7bf62caa091a82282096d19c04e717ec50a178f20c"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.147157 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.147166 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" event={"ID":"f60d64f9-1517-4093-9a99-abd333d12a36","Type":"ContainerStarted","Data":"59e7e33d832f61a2e3f267769ce208f515c2fad8fff6af6b97cccd7206e3944a"} Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.193870 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.194977 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.198804 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:44:28 crc kubenswrapper[4836]: E0122 09:44:28.223777 4836 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf699f3_dacd_4f97_97b3_0bb300377623.slice\": RecentStats: unable to find data in memory cache]" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.241438 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.241487 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.242686 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.302620 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.346136 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.346184 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.346222 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb99t\" (UniqueName: \"kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.359472 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.447636 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.447681 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.447724 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb99t\" (UniqueName: \"kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.448348 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.448488 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.454378 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" podStartSLOduration=132.454359993 podStartE2EDuration="2m12.454359993s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:28.315661825 +0000 UTC m=+151.675430342" watchObservedRunningTime="2026-01-22 09:44:28.454359993 +0000 UTC m=+151.814128510" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.454921 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.469895 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb99t\" (UniqueName: \"kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t\") pod \"redhat-marketplace-86vhq\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: W0122 09:44:28.482824 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89c88a4_5375_47f6_ab66_07aa49a11a2e.slice/crio-c696a1844cb19bd078e0ee2e9c34d50c9f419109328e5a8bffa2cfaa5da39653 WatchSource:0}: Error finding container c696a1844cb19bd078e0ee2e9c34d50c9f419109328e5a8bffa2cfaa5da39653: Status 404 returned error can't find the container with id c696a1844cb19bd078e0ee2e9c34d50c9f419109328e5a8bffa2cfaa5da39653 Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.522861 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.644017 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 22 09:44:28 crc kubenswrapper[4836]: I0122 09:44:28.776668 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:44:28 crc kubenswrapper[4836]: W0122 09:44:28.782161 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4969f850_dcc4_4ebf_b3cf_9a59e7f8f0d0.slice/crio-f2d91f91b5ab938811a79ec35bb1290582519eab5ea5fbd5f40ca2efe589a2b8 WatchSource:0}: Error finding container f2d91f91b5ab938811a79ec35bb1290582519eab5ea5fbd5f40ca2efe589a2b8: Status 404 returned error can't find the container with id f2d91f91b5ab938811a79ec35bb1290582519eab5ea5fbd5f40ca2efe589a2b8 Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.064924 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:29 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:29 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:29 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.065015 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.153353 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerStarted","Data":"c696a1844cb19bd078e0ee2e9c34d50c9f419109328e5a8bffa2cfaa5da39653"} Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.155375 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c","Type":"ContainerStarted","Data":"95c57b70284f169de81e4e9a57637678988e5791978963047d9982c0ce67a134"} Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.156446 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerStarted","Data":"f2d91f91b5ab938811a79ec35bb1290582519eab5ea5fbd5f40ca2efe589a2b8"} Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.182717 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.184485 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.186905 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.195032 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.260470 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.260516 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jws8b\" (UniqueName: \"kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.260860 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.366244 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.367566 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.367601 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jws8b\" (UniqueName: \"kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.366794 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.367985 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.392170 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jws8b\" (UniqueName: \"kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b\") pod \"redhat-operators-cpx6h\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.478114 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.526644 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.572574 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir\") pod \"986b7f3a-87ee-473f-926b-104b896b7d31\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.572807 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access\") pod \"986b7f3a-87ee-473f-926b-104b896b7d31\" (UID: \"986b7f3a-87ee-473f-926b-104b896b7d31\") " Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.572986 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "986b7f3a-87ee-473f-926b-104b896b7d31" (UID: "986b7f3a-87ee-473f-926b-104b896b7d31"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.573468 4836 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/986b7f3a-87ee-473f-926b-104b896b7d31-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.578841 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "986b7f3a-87ee-473f-926b-104b896b7d31" (UID: "986b7f3a-87ee-473f-926b-104b896b7d31"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.580835 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:44:29 crc kubenswrapper[4836]: E0122 09:44:29.581068 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986b7f3a-87ee-473f-926b-104b896b7d31" containerName="pruner" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.581081 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="986b7f3a-87ee-473f-926b-104b896b7d31" containerName="pruner" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.581170 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="986b7f3a-87ee-473f-926b-104b896b7d31" containerName="pruner" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.581826 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.590675 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.675119 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nwsz\" (UniqueName: \"kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.675220 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.675278 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.675380 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/986b7f3a-87ee-473f-926b-104b896b7d31-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.696959 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.776719 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.776782 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.777098 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nwsz\" (UniqueName: \"kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.777548 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.777594 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.795442 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nwsz\" (UniqueName: \"kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz\") pod \"redhat-operators-2xnzx\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:29 crc kubenswrapper[4836]: I0122 09:44:29.908435 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.063139 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:30 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:30 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:30 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.063232 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.163433 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c","Type":"ContainerStarted","Data":"ae61121e131427d11fe44dfb39599e894ac2a24107b93c66cac3261ab31af2ca"} Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.165306 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"986b7f3a-87ee-473f-926b-104b896b7d31","Type":"ContainerDied","Data":"b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4"} Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.165333 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.165339 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0001580460b0e0c2eafb92da69dfe173520a650b097dce2a5c91734b4590eb4" Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.166664 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerStarted","Data":"fb304af5c031253133d5b8b38bd1753f8c670fbdb2b5f0f924f3a99a12633edf"} Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.168425 4836 generic.go:334] "Generic (PLEG): container finished" podID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerID="aebb3b8fd635bcbc64d2ba7a6d6727c22de7245cb4231d8e8104b4ff6be8a4f3" exitCode=0 Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.168539 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerDied","Data":"aebb3b8fd635bcbc64d2ba7a6d6727c22de7245cb4231d8e8104b4ff6be8a4f3"} Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.176157 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerStarted","Data":"55be1a33fa4ebccbe9e46fdb32f4de6b6136879f08315b21a4e95eeb34d1997e"} Jan 22 09:44:30 crc kubenswrapper[4836]: I0122 09:44:30.440323 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:44:30 crc kubenswrapper[4836]: W0122 09:44:30.443604 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69209694_e7d1_4e15_bb51_2807be65ed0a.slice/crio-473a1d29c9c6fe49dac4e0475a7e2f09aa00df5d0a6d7fa0e7ae86566dcf688d WatchSource:0}: Error finding container 473a1d29c9c6fe49dac4e0475a7e2f09aa00df5d0a6d7fa0e7ae86566dcf688d: Status 404 returned error can't find the container with id 473a1d29c9c6fe49dac4e0475a7e2f09aa00df5d0a6d7fa0e7ae86566dcf688d Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.062061 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:31 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:31 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:31 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.062348 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.181446 4836 generic.go:334] "Generic (PLEG): container finished" podID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerID="fb304af5c031253133d5b8b38bd1753f8c670fbdb2b5f0f924f3a99a12633edf" exitCode=0 Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.181525 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerDied","Data":"fb304af5c031253133d5b8b38bd1753f8c670fbdb2b5f0f924f3a99a12633edf"} Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.183203 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerStarted","Data":"473a1d29c9c6fe49dac4e0475a7e2f09aa00df5d0a6d7fa0e7ae86566dcf688d"} Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.680168 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.680211 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:31 crc kubenswrapper[4836]: I0122 09:44:31.686446 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:32 crc kubenswrapper[4836]: I0122 09:44:32.062736 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:32 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:32 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:32 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:32 crc kubenswrapper[4836]: I0122 09:44:32.062831 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:32 crc kubenswrapper[4836]: I0122 09:44:32.196584 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-txqw8" Jan 22 09:44:33 crc kubenswrapper[4836]: I0122 09:44:33.062174 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:33 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:33 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:33 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:33 crc kubenswrapper[4836]: I0122 09:44:33.062569 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:33 crc kubenswrapper[4836]: I0122 09:44:33.216492 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=5.216462046 podStartE2EDuration="5.216462046s" podCreationTimestamp="2026-01-22 09:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:33.213581605 +0000 UTC m=+156.573350132" watchObservedRunningTime="2026-01-22 09:44:33.216462046 +0000 UTC m=+156.576230573" Jan 22 09:44:33 crc kubenswrapper[4836]: I0122 09:44:33.925315 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-c7hkn" Jan 22 09:44:34 crc kubenswrapper[4836]: I0122 09:44:34.061629 4836 patch_prober.go:28] interesting pod/router-default-5444994796-9mq8w container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 22 09:44:34 crc kubenswrapper[4836]: [-]has-synced failed: reason withheld Jan 22 09:44:34 crc kubenswrapper[4836]: [+]process-running ok Jan 22 09:44:34 crc kubenswrapper[4836]: healthz check failed Jan 22 09:44:34 crc kubenswrapper[4836]: I0122 09:44:34.061690 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9mq8w" podUID="5afa5ef0-f947-4bf6-82b2-99b47e25e10d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.063928 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.071360 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9mq8w" Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.081761 4836 patch_prober.go:28] interesting pod/console-f9d7485db-tbm5x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.081804 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-tbm5x" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.216092 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerStarted","Data":"15bd7ceb1ed697fafeddee904695f4d1b182a3c54c84cffac30645dae1b0fbff"} Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.222755 4836 generic.go:334] "Generic (PLEG): container finished" podID="eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" containerID="ae61121e131427d11fe44dfb39599e894ac2a24107b93c66cac3261ab31af2ca" exitCode=0 Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.223534 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c","Type":"ContainerDied","Data":"ae61121e131427d11fe44dfb39599e894ac2a24107b93c66cac3261ab31af2ca"} Jan 22 09:44:35 crc kubenswrapper[4836]: I0122 09:44:35.713963 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-z9w5n" Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.229654 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerStarted","Data":"8d5d71d53566bb003146644f8b6e4078ab590c33329b29b061308206ec9b567a"} Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.488499 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.582731 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir\") pod \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.582830 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access\") pod \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\" (UID: \"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c\") " Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.582870 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" (UID: "eb84fc20-5de4-4e5a-bb2c-1cdce507a27c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.583110 4836 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.604353 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" (UID: "eb84fc20-5de4-4e5a-bb2c-1cdce507a27c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:44:36 crc kubenswrapper[4836]: I0122 09:44:36.683971 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb84fc20-5de4-4e5a-bb2c-1cdce507a27c-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.255215 4836 generic.go:334] "Generic (PLEG): container finished" podID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerID="8d5d71d53566bb003146644f8b6e4078ab590c33329b29b061308206ec9b567a" exitCode=0 Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.255313 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerDied","Data":"8d5d71d53566bb003146644f8b6e4078ab590c33329b29b061308206ec9b567a"} Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.257602 4836 generic.go:334] "Generic (PLEG): container finished" podID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerID="15bd7ceb1ed697fafeddee904695f4d1b182a3c54c84cffac30645dae1b0fbff" exitCode=0 Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.257672 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerDied","Data":"15bd7ceb1ed697fafeddee904695f4d1b182a3c54c84cffac30645dae1b0fbff"} Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.259945 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb84fc20-5de4-4e5a-bb2c-1cdce507a27c","Type":"ContainerDied","Data":"95c57b70284f169de81e4e9a57637678988e5791978963047d9982c0ce67a134"} Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.259976 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95c57b70284f169de81e4e9a57637678988e5791978963047d9982c0ce67a134" Jan 22 09:44:37 crc kubenswrapper[4836]: I0122 09:44:37.260023 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.513895 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.520164 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ab725c3-0744-400a-b8d7-66e1165476e5-metrics-certs\") pod \"network-metrics-daemon-6kkmw\" (UID: \"9ab725c3-0744-400a-b8d7-66e1165476e5\") " pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.663263 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.665717 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" containerID="cri-o://63b6238ad7a62b56c78627f4021f1b06eefec5ec90a01d055f76f76de6bfd38c" gracePeriod=30 Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.674517 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.674762 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" containerID="cri-o://4455a6b2e0b04552492f4b39c68ccb873bd23797ca9c94f3a9542e36e40c6fbe" gracePeriod=30 Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.713860 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6kkmw" Jan 22 09:44:38 crc kubenswrapper[4836]: I0122 09:44:38.901871 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6kkmw"] Jan 22 09:44:38 crc kubenswrapper[4836]: W0122 09:44:38.907513 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ab725c3_0744_400a_b8d7_66e1165476e5.slice/crio-ce3abbc2e84d756123e777616723c1ec05c438250a4bf9279140d9508013cf92 WatchSource:0}: Error finding container ce3abbc2e84d756123e777616723c1ec05c438250a4bf9279140d9508013cf92: Status 404 returned error can't find the container with id ce3abbc2e84d756123e777616723c1ec05c438250a4bf9279140d9508013cf92 Jan 22 09:44:39 crc kubenswrapper[4836]: I0122 09:44:39.275355 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" event={"ID":"9ab725c3-0744-400a-b8d7-66e1165476e5","Type":"ContainerStarted","Data":"ce3abbc2e84d756123e777616723c1ec05c438250a4bf9279140d9508013cf92"} Jan 22 09:44:39 crc kubenswrapper[4836]: I0122 09:44:39.277631 4836 generic.go:334] "Generic (PLEG): container finished" podID="e66435af-b690-40a3-a039-63a541178060" containerID="4455a6b2e0b04552492f4b39c68ccb873bd23797ca9c94f3a9542e36e40c6fbe" exitCode=0 Jan 22 09:44:39 crc kubenswrapper[4836]: I0122 09:44:39.277666 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" event={"ID":"e66435af-b690-40a3-a039-63a541178060","Type":"ContainerDied","Data":"4455a6b2e0b04552492f4b39c68ccb873bd23797ca9c94f3a9542e36e40c6fbe"} Jan 22 09:44:40 crc kubenswrapper[4836]: I0122 09:44:40.284292 4836 generic.go:334] "Generic (PLEG): container finished" podID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerID="63b6238ad7a62b56c78627f4021f1b06eefec5ec90a01d055f76f76de6bfd38c" exitCode=0 Jan 22 09:44:40 crc kubenswrapper[4836]: I0122 09:44:40.284347 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" event={"ID":"5148a0f2-8687-4dd0-af2c-58cd45924d13","Type":"ContainerDied","Data":"63b6238ad7a62b56c78627f4021f1b06eefec5ec90a01d055f76f76de6bfd38c"} Jan 22 09:44:44 crc kubenswrapper[4836]: I0122 09:44:44.311914 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" event={"ID":"9ab725c3-0744-400a-b8d7-66e1165476e5","Type":"ContainerStarted","Data":"9d1b9cb7a215d2cb1525dd187889054be68e1127bd81491c86e7b794d9959bb7"} Jan 22 09:44:44 crc kubenswrapper[4836]: I0122 09:44:44.853316 4836 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-jwkz2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:44:44 crc kubenswrapper[4836]: I0122 09:44:44.853406 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 22 09:44:45 crc kubenswrapper[4836]: I0122 09:44:45.084966 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:45 crc kubenswrapper[4836]: I0122 09:44:45.089543 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:44:46 crc kubenswrapper[4836]: I0122 09:44:46.579871 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:44:46 crc kubenswrapper[4836]: I0122 09:44:46.986339 4836 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-m66s2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:44:46 crc kubenswrapper[4836]: I0122 09:44:46.986684 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 09:44:47 crc kubenswrapper[4836]: I0122 09:44:47.663648 4836 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mfxbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:44:47 crc kubenswrapper[4836]: I0122 09:44:47.663768 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 09:44:51 crc kubenswrapper[4836]: I0122 09:44:51.131094 4836 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.059270645s: [/var/lib/containers/storage/overlay/2da02de10171455666786834c40deb81e3c10257e3ef23caffab53b925fae16f/diff /var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/4.log]; will not log again for this container unless duration exceeds 2s Jan 22 09:44:53 crc kubenswrapper[4836]: I0122 09:44:53.897384 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:44:53 crc kubenswrapper[4836]: I0122 09:44:53.897462 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.597722 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.634283 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:44:55 crc kubenswrapper[4836]: E0122 09:44:55.634782 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.634804 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" Jan 22 09:44:55 crc kubenswrapper[4836]: E0122 09:44:55.634839 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" containerName="pruner" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.634848 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" containerName="pruner" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.635094 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" containerName="route-controller-manager" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.635119 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb84fc20-5de4-4e5a-bb2c-1cdce507a27c" containerName="pruner" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.635781 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.651977 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.655652 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca\") pod \"5148a0f2-8687-4dd0-af2c-58cd45924d13\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.655697 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert\") pod \"5148a0f2-8687-4dd0-af2c-58cd45924d13\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.655745 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config\") pod \"5148a0f2-8687-4dd0-af2c-58cd45924d13\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.655795 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8szz\" (UniqueName: \"kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz\") pod \"5148a0f2-8687-4dd0-af2c-58cd45924d13\" (UID: \"5148a0f2-8687-4dd0-af2c-58cd45924d13\") " Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.660261 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config" (OuterVolumeSpecName: "config") pod "5148a0f2-8687-4dd0-af2c-58cd45924d13" (UID: "5148a0f2-8687-4dd0-af2c-58cd45924d13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.662361 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca" (OuterVolumeSpecName: "client-ca") pod "5148a0f2-8687-4dd0-af2c-58cd45924d13" (UID: "5148a0f2-8687-4dd0-af2c-58cd45924d13"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.668521 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5148a0f2-8687-4dd0-af2c-58cd45924d13" (UID: "5148a0f2-8687-4dd0-af2c-58cd45924d13"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.668599 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz" (OuterVolumeSpecName: "kube-api-access-t8szz") pod "5148a0f2-8687-4dd0-af2c-58cd45924d13" (UID: "5148a0f2-8687-4dd0-af2c-58cd45924d13"). InnerVolumeSpecName "kube-api-access-t8szz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.761801 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.761850 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.761874 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kbdq\" (UniqueName: \"kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.761925 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.761997 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.762009 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5148a0f2-8687-4dd0-af2c-58cd45924d13-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.762017 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5148a0f2-8687-4dd0-af2c-58cd45924d13-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.762025 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8szz\" (UniqueName: \"kubernetes.io/projected/5148a0f2-8687-4dd0-af2c-58cd45924d13-kube-api-access-t8szz\") on node \"crc\" DevicePath \"\"" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.862688 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.862739 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.862765 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kbdq\" (UniqueName: \"kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.863282 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.864213 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.864279 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.866772 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.887057 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kbdq\" (UniqueName: \"kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq\") pod \"route-controller-manager-5d46f96776-sdcp2\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:55 crc kubenswrapper[4836]: I0122 09:44:55.969898 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.192136 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-n6njz" Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.383080 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" event={"ID":"5148a0f2-8687-4dd0-af2c-58cd45924d13","Type":"ContainerDied","Data":"bba508c3fb7b38ccfd628a9586f625afe5c348f822d5013cb17c52550be57161"} Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.383152 4836 scope.go:117] "RemoveContainer" containerID="63b6238ad7a62b56c78627f4021f1b06eefec5ec90a01d055f76f76de6bfd38c" Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.383158 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2" Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.410943 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:56 crc kubenswrapper[4836]: I0122 09:44:56.413334 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m66s2"] Jan 22 09:44:57 crc kubenswrapper[4836]: I0122 09:44:57.005990 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5148a0f2-8687-4dd0-af2c-58cd45924d13" path="/var/lib/kubelet/pods/5148a0f2-8687-4dd0-af2c-58cd45924d13/volumes" Jan 22 09:44:57 crc kubenswrapper[4836]: I0122 09:44:57.663829 4836 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mfxbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:44:57 crc kubenswrapper[4836]: I0122 09:44:57.663892 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 09:44:58 crc kubenswrapper[4836]: I0122 09:44:58.440101 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.137735 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp"] Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.138946 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.141179 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.141326 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.146559 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp"] Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.222434 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.222521 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.222561 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x2jk\" (UniqueName: \"kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.324094 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.324724 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.324791 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x2jk\" (UniqueName: \"kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.326594 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.329980 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.341005 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x2jk\" (UniqueName: \"kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk\") pod \"collect-profiles-29484585-v4rvp\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:00 crc kubenswrapper[4836]: I0122 09:45:00.454522 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.537628 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.539089 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.545073 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.553962 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.555516 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.692760 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.693035 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.795186 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.795290 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.795354 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.828969 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:04 crc kubenswrapper[4836]: I0122 09:45:04.867953 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:06 crc kubenswrapper[4836]: I0122 09:45:06.804482 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 22 09:45:07 crc kubenswrapper[4836]: I0122 09:45:07.668032 4836 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-mfxbj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 22 09:45:07 crc kubenswrapper[4836]: I0122 09:45:07.668359 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.709950 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.740689 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:45:09 crc kubenswrapper[4836]: E0122 09:45:09.740902 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.740913 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.741018 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="e66435af-b690-40a3-a039-63a541178060" containerName="controller-manager" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.741839 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.751326 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.859638 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") pod \"e66435af-b690-40a3-a039-63a541178060\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.859690 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") pod \"e66435af-b690-40a3-a039-63a541178060\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.859734 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") pod \"e66435af-b690-40a3-a039-63a541178060\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.859779 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") pod \"e66435af-b690-40a3-a039-63a541178060\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.859839 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") pod \"e66435af-b690-40a3-a039-63a541178060\" (UID: \"e66435af-b690-40a3-a039-63a541178060\") " Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860077 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860127 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860162 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx8wj\" (UniqueName: \"kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860188 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860478 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.860624 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca" (OuterVolumeSpecName: "client-ca") pod "e66435af-b690-40a3-a039-63a541178060" (UID: "e66435af-b690-40a3-a039-63a541178060"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.861256 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e66435af-b690-40a3-a039-63a541178060" (UID: "e66435af-b690-40a3-a039-63a541178060"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.861438 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config" (OuterVolumeSpecName: "config") pod "e66435af-b690-40a3-a039-63a541178060" (UID: "e66435af-b690-40a3-a039-63a541178060"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.865291 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd" (OuterVolumeSpecName: "kube-api-access-v68xd") pod "e66435af-b690-40a3-a039-63a541178060" (UID: "e66435af-b690-40a3-a039-63a541178060"). InnerVolumeSpecName "kube-api-access-v68xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.865755 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e66435af-b690-40a3-a039-63a541178060" (UID: "e66435af-b690-40a3-a039-63a541178060"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.961844 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.961921 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.961954 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.961975 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx8wj\" (UniqueName: \"kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962000 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962062 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962073 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v68xd\" (UniqueName: \"kubernetes.io/projected/e66435af-b690-40a3-a039-63a541178060-kube-api-access-v68xd\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962084 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962093 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e66435af-b690-40a3-a039-63a541178060-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.962100 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e66435af-b690-40a3-a039-63a541178060-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.963305 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.963470 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.964137 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.966306 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:09 crc kubenswrapper[4836]: I0122 09:45:09.978878 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx8wj\" (UniqueName: \"kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj\") pod \"controller-manager-6ff6dd6f76-llz2b\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.063503 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.132055 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.132972 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.144764 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.265346 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.265429 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.265471 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.366973 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.367068 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.367108 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.367211 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.367205 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.384245 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.455858 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.463431 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" event={"ID":"e66435af-b690-40a3-a039-63a541178060","Type":"ContainerDied","Data":"d15eebeacd158a8e66f082cecda80d5930ccabc5e827b892c7fd0c28183a744a"} Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.463503 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-mfxbj" Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.492240 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:45:10 crc kubenswrapper[4836]: I0122 09:45:10.494951 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-mfxbj"] Jan 22 09:45:11 crc kubenswrapper[4836]: I0122 09:45:11.006148 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e66435af-b690-40a3-a039-63a541178060" path="/var/lib/kubelet/pods/e66435af-b690-40a3-a039-63a541178060/volumes" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.806663 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\": context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.807393 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nwsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2xnzx_openshift-marketplace(69209694-e7d1-4e15-bb51-2807be65ed0a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\": context canceled" logger="UnhandledError" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.808680 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \\\"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\\\": context canceled\"" pod="openshift-marketplace/redhat-operators-2xnzx" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.834851 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\": context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.835255 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jws8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cpx6h_openshift-marketplace(4363607a-4a5d-4e80-9389-4cd4214e316b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\": context canceled" logger="UnhandledError" Jan 22 09:45:20 crc kubenswrapper[4836]: E0122 09:45:20.836525 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea: Get \\\"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:375463ce314e9870c2ef316f6ae8ec2bad821721d7dac5d2800db42bce264bea\\\": context canceled\"" pod="openshift-marketplace/redhat-operators-cpx6h" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" Jan 22 09:45:20 crc kubenswrapper[4836]: I0122 09:45:20.841518 4836 scope.go:117] "RemoveContainer" containerID="4455a6b2e0b04552492f4b39c68ccb873bd23797ca9c94f3a9542e36e40c6fbe" Jan 22 09:45:22 crc kubenswrapper[4836]: E0122 09:45:22.640811 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 09:45:22 crc kubenswrapper[4836]: E0122 09:45:22.641592 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9m69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ztdcc_openshift-marketplace(f89c88a4-5375-47f6-ab66-07aa49a11a2e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:22 crc kubenswrapper[4836]: E0122 09:45:22.642989 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ztdcc" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" Jan 22 09:45:23 crc kubenswrapper[4836]: I0122 09:45:23.896926 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:45:23 crc kubenswrapper[4836]: I0122 09:45:23.897162 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:45:23 crc kubenswrapper[4836]: I0122 09:45:23.897204 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:45:23 crc kubenswrapper[4836]: I0122 09:45:23.897696 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 09:45:23 crc kubenswrapper[4836]: I0122 09:45:23.897776 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f" gracePeriod=600 Jan 22 09:45:23 crc kubenswrapper[4836]: E0122 09:45:23.919915 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ztdcc" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" Jan 22 09:45:23 crc kubenswrapper[4836]: E0122 09:45:23.979074 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 09:45:23 crc kubenswrapper[4836]: E0122 09:45:23.979503 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbnjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zbvh5_openshift-marketplace(9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:23 crc kubenswrapper[4836]: E0122 09:45:23.980651 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zbvh5" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" Jan 22 09:45:24 crc kubenswrapper[4836]: I0122 09:45:24.545714 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f" exitCode=0 Jan 22 09:45:24 crc kubenswrapper[4836]: I0122 09:45:24.546168 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f"} Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.724767 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zbvh5" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.819258 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.819767 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2crrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6pq7n_openshift-marketplace(d7de3ebe-e601-46b3-bda0-8e08d0a42e43): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.820944 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6pq7n" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.834939 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.835106 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfr6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7bfkp_openshift-marketplace(65b61041-36cc-4c4d-be61-dd4e370b7ff3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.836987 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7bfkp" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.843360 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.843539 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hb99t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-86vhq_openshift-marketplace(4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.845244 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-86vhq" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.990535 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.990990 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvznw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-n4dwm_openshift-marketplace(1b2b98a3-9481-442f-a35a-d8adcf779887): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:25 crc kubenswrapper[4836]: E0122 09:45:25.992252 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-n4dwm" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.264278 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 22 09:45:26 crc kubenswrapper[4836]: W0122 09:45:26.274326 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1c7b55b5_32c4_404a_a3f4_e56a6622e581.slice/crio-af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab WatchSource:0}: Error finding container af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab: Status 404 returned error can't find the container with id af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.279601 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:45:26 crc kubenswrapper[4836]: W0122 09:45:26.291013 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66594c84_2142_4309_9c8e_ae1aaf196d7d.slice/crio-ba901edb4ac5fd079475cf51b02f92102b9ef34b8c58614eb8844c62564d7487 WatchSource:0}: Error finding container ba901edb4ac5fd079475cf51b02f92102b9ef34b8c58614eb8844c62564d7487: Status 404 returned error can't find the container with id ba901edb4ac5fd079475cf51b02f92102b9ef34b8c58614eb8844c62564d7487 Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.346323 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp"] Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.353208 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.364298 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.559906 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63b6b777-0f50-4dd2-af16-1f27e8969a7f","Type":"ContainerStarted","Data":"e9e08e92b548f6c42cdb64fb44e9c2e02381457938959655a3c78ef739e99d80"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.561798 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" event={"ID":"c1b43b9c-abd0-4c0f-91d5-d92477b97487","Type":"ContainerStarted","Data":"9ac3b106a22c56e70052efa733be76c99dd7275a87c0cb70735aa4fa18206d27"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.562139 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" event={"ID":"c1b43b9c-abd0-4c0f-91d5-d92477b97487","Type":"ContainerStarted","Data":"d80cd8f8377ef09a034fd7203fa230865a9c42f80809d75c44be17de8feb8b06"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.565171 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6kkmw" event={"ID":"9ab725c3-0744-400a-b8d7-66e1165476e5","Type":"ContainerStarted","Data":"bab9541804be5cb216956fd2e0bb0378cb4e0e265d38995f7dfec062b1123c58"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.567674 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" event={"ID":"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656","Type":"ContainerStarted","Data":"6764cfdd6a9cc843c1e37806a36d6130c3e7ec7f88acf9415cdee2caec80e519"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.567710 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" event={"ID":"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656","Type":"ContainerStarted","Data":"3c82540160b00998f6b4c0fd7c100efdbc40d6c935e6dfcfdceda02a16f535f9"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.568481 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.569909 4836 patch_prober.go:28] interesting pod/controller-manager-6ff6dd6f76-llz2b container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.569944 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.573362 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"1c7b55b5-32c4-404a-a3f4-e56a6622e581","Type":"ContainerStarted","Data":"af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.577218 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" event={"ID":"66594c84-2142-4309-9c8e-ae1aaf196d7d","Type":"ContainerStarted","Data":"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.577464 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" event={"ID":"66594c84-2142-4309-9c8e-ae1aaf196d7d","Type":"ContainerStarted","Data":"ba901edb4ac5fd079475cf51b02f92102b9ef34b8c58614eb8844c62564d7487"} Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.577475 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerName="route-controller-manager" containerID="cri-o://390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68" gracePeriod=30 Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.577565 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.578747 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" podStartSLOduration=26.578725404 podStartE2EDuration="26.578725404s" podCreationTimestamp="2026-01-22 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:26.576690641 +0000 UTC m=+209.936459188" watchObservedRunningTime="2026-01-22 09:45:26.578725404 +0000 UTC m=+209.938493921" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.581734 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86"} Jan 22 09:45:26 crc kubenswrapper[4836]: E0122 09:45:26.590468 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-86vhq" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" Jan 22 09:45:26 crc kubenswrapper[4836]: E0122 09:45:26.593177 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6pq7n" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" Jan 22 09:45:26 crc kubenswrapper[4836]: E0122 09:45:26.593377 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-n4dwm" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.599069 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6kkmw" podStartSLOduration=190.599048989 podStartE2EDuration="3m10.599048989s" podCreationTimestamp="2026-01-22 09:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:26.591736803 +0000 UTC m=+209.951505320" watchObservedRunningTime="2026-01-22 09:45:26.599048989 +0000 UTC m=+209.958817506" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.636964 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" podStartSLOduration=28.636949533 podStartE2EDuration="28.636949533s" podCreationTimestamp="2026-01-22 09:44:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:26.614365649 +0000 UTC m=+209.974134166" watchObservedRunningTime="2026-01-22 09:45:26.636949533 +0000 UTC m=+209.996718050" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.675434 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" podStartSLOduration=48.675415233 podStartE2EDuration="48.675415233s" podCreationTimestamp="2026-01-22 09:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:26.652338896 +0000 UTC m=+210.012107433" watchObservedRunningTime="2026-01-22 09:45:26.675415233 +0000 UTC m=+210.035183750" Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.922532 4836 patch_prober.go:28] interesting pod/route-controller-manager-5d46f96776-sdcp2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:41504->10.217.0.54:8443: read: connection reset by peer" start-of-body= Jan 22 09:45:26 crc kubenswrapper[4836]: I0122 09:45:26.922932 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:41504->10.217.0.54:8443: read: connection reset by peer" Jan 22 09:45:27 crc kubenswrapper[4836]: I0122 09:45:27.595931 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"1c7b55b5-32c4-404a-a3f4-e56a6622e581","Type":"ContainerStarted","Data":"e1d07371fc3c5199caeb61d8bdbed72fa1daf640cc825d7e3f2db3ac33aa2c01"} Jan 22 09:45:27 crc kubenswrapper[4836]: I0122 09:45:27.597593 4836 generic.go:334] "Generic (PLEG): container finished" podID="c1b43b9c-abd0-4c0f-91d5-d92477b97487" containerID="9ac3b106a22c56e70052efa733be76c99dd7275a87c0cb70735aa4fa18206d27" exitCode=0 Jan 22 09:45:27 crc kubenswrapper[4836]: I0122 09:45:27.597682 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" event={"ID":"c1b43b9c-abd0-4c0f-91d5-d92477b97487","Type":"ContainerDied","Data":"9ac3b106a22c56e70052efa733be76c99dd7275a87c0cb70735aa4fa18206d27"} Jan 22 09:45:27 crc kubenswrapper[4836]: I0122 09:45:27.603176 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:27 crc kubenswrapper[4836]: I0122 09:45:27.613014 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=23.612986449 podStartE2EDuration="23.612986449s" podCreationTimestamp="2026-01-22 09:45:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:27.61264876 +0000 UTC m=+210.972417287" watchObservedRunningTime="2026-01-22 09:45:27.612986449 +0000 UTC m=+210.972754986" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.576349 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-5d46f96776-sdcp2_66594c84-2142-4309-9c8e-ae1aaf196d7d/route-controller-manager/0.log" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.576862 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.605267 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:45:28 crc kubenswrapper[4836]: E0122 09:45:28.605562 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerName="route-controller-manager" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.605587 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerName="route-controller-manager" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.605731 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerName="route-controller-manager" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.606192 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.615367 4836 generic.go:334] "Generic (PLEG): container finished" podID="1c7b55b5-32c4-404a-a3f4-e56a6622e581" containerID="e1d07371fc3c5199caeb61d8bdbed72fa1daf640cc825d7e3f2db3ac33aa2c01" exitCode=0 Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.615434 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"1c7b55b5-32c4-404a-a3f4-e56a6622e581","Type":"ContainerDied","Data":"e1d07371fc3c5199caeb61d8bdbed72fa1daf640cc825d7e3f2db3ac33aa2c01"} Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621438 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-5d46f96776-sdcp2_66594c84-2142-4309-9c8e-ae1aaf196d7d/route-controller-manager/0.log" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621499 4836 generic.go:334] "Generic (PLEG): container finished" podID="66594c84-2142-4309-9c8e-ae1aaf196d7d" containerID="390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68" exitCode=255 Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621586 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621607 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" event={"ID":"66594c84-2142-4309-9c8e-ae1aaf196d7d","Type":"ContainerDied","Data":"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68"} Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621718 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2" event={"ID":"66594c84-2142-4309-9c8e-ae1aaf196d7d","Type":"ContainerDied","Data":"ba901edb4ac5fd079475cf51b02f92102b9ef34b8c58614eb8844c62564d7487"} Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.621741 4836 scope.go:117] "RemoveContainer" containerID="390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.625567 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63b6b777-0f50-4dd2-af16-1f27e8969a7f","Type":"ContainerStarted","Data":"d491c6962735d4108d8e22f896d7185b9f600798596cf71691ab6418042bf9f8"} Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.625935 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.648767 4836 scope.go:117] "RemoveContainer" containerID="390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68" Jan 22 09:45:28 crc kubenswrapper[4836]: E0122 09:45:28.650820 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68\": container with ID starting with 390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68 not found: ID does not exist" containerID="390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.650886 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68"} err="failed to get container status \"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68\": rpc error: code = NotFound desc = could not find container \"390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68\": container with ID starting with 390401c64e15bf122857fd407ebe08171655e6d9c0c5abe40c583c0c8ce8fe68 not found: ID does not exist" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.682700 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=18.682667952 podStartE2EDuration="18.682667952s" podCreationTimestamp="2026-01-22 09:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:28.674568506 +0000 UTC m=+212.034337023" watchObservedRunningTime="2026-01-22 09:45:28.682667952 +0000 UTC m=+212.042436469" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735331 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kbdq\" (UniqueName: \"kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq\") pod \"66594c84-2142-4309-9c8e-ae1aaf196d7d\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735550 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config\") pod \"66594c84-2142-4309-9c8e-ae1aaf196d7d\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735568 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca\") pod \"66594c84-2142-4309-9c8e-ae1aaf196d7d\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735602 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert\") pod \"66594c84-2142-4309-9c8e-ae1aaf196d7d\" (UID: \"66594c84-2142-4309-9c8e-ae1aaf196d7d\") " Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735867 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735891 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735955 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.735985 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d5j5\" (UniqueName: \"kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.737296 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca" (OuterVolumeSpecName: "client-ca") pod "66594c84-2142-4309-9c8e-ae1aaf196d7d" (UID: "66594c84-2142-4309-9c8e-ae1aaf196d7d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.737980 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config" (OuterVolumeSpecName: "config") pod "66594c84-2142-4309-9c8e-ae1aaf196d7d" (UID: "66594c84-2142-4309-9c8e-ae1aaf196d7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.745300 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "66594c84-2142-4309-9c8e-ae1aaf196d7d" (UID: "66594c84-2142-4309-9c8e-ae1aaf196d7d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.745327 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq" (OuterVolumeSpecName: "kube-api-access-2kbdq") pod "66594c84-2142-4309-9c8e-ae1aaf196d7d" (UID: "66594c84-2142-4309-9c8e-ae1aaf196d7d"). InnerVolumeSpecName "kube-api-access-2kbdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.837290 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.837922 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838029 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838240 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d5j5\" (UniqueName: \"kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838370 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kbdq\" (UniqueName: \"kubernetes.io/projected/66594c84-2142-4309-9c8e-ae1aaf196d7d-kube-api-access-2kbdq\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838440 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838498 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66594c84-2142-4309-9c8e-ae1aaf196d7d-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.838553 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66594c84-2142-4309-9c8e-ae1aaf196d7d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.841078 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.842296 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.847186 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.856738 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d5j5\" (UniqueName: \"kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5\") pod \"route-controller-manager-6995bc88c8-4nq46\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.895630 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.923922 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.958519 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:45:28 crc kubenswrapper[4836]: I0122 09:45:28.961843 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d46f96776-sdcp2"] Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.008204 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66594c84-2142-4309-9c8e-ae1aaf196d7d" path="/var/lib/kubelet/pods/66594c84-2142-4309-9c8e-ae1aaf196d7d/volumes" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.045122 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x2jk\" (UniqueName: \"kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk\") pod \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.045535 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume\") pod \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.045658 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume\") pod \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\" (UID: \"c1b43b9c-abd0-4c0f-91d5-d92477b97487\") " Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.046087 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume" (OuterVolumeSpecName: "config-volume") pod "c1b43b9c-abd0-4c0f-91d5-d92477b97487" (UID: "c1b43b9c-abd0-4c0f-91d5-d92477b97487"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.048510 4836 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c1b43b9c-abd0-4c0f-91d5-d92477b97487-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.049853 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c1b43b9c-abd0-4c0f-91d5-d92477b97487" (UID: "c1b43b9c-abd0-4c0f-91d5-d92477b97487"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.050572 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk" (OuterVolumeSpecName: "kube-api-access-2x2jk") pod "c1b43b9c-abd0-4c0f-91d5-d92477b97487" (UID: "c1b43b9c-abd0-4c0f-91d5-d92477b97487"). InnerVolumeSpecName "kube-api-access-2x2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.150484 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x2jk\" (UniqueName: \"kubernetes.io/projected/c1b43b9c-abd0-4c0f-91d5-d92477b97487-kube-api-access-2x2jk\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.150836 4836 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c1b43b9c-abd0-4c0f-91d5-d92477b97487-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.329456 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:45:29 crc kubenswrapper[4836]: W0122 09:45:29.338586 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e3bf656_aa9a_4e7c_8d80_78be8d462abb.slice/crio-86b2911edec33b3ad83ad0811f8c7c1a165b335612c5cddb5ed738e8ce2fe427 WatchSource:0}: Error finding container 86b2911edec33b3ad83ad0811f8c7c1a165b335612c5cddb5ed738e8ce2fe427: Status 404 returned error can't find the container with id 86b2911edec33b3ad83ad0811f8c7c1a165b335612c5cddb5ed738e8ce2fe427 Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.632750 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" event={"ID":"c1b43b9c-abd0-4c0f-91d5-d92477b97487","Type":"ContainerDied","Data":"d80cd8f8377ef09a034fd7203fa230865a9c42f80809d75c44be17de8feb8b06"} Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.633100 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d80cd8f8377ef09a034fd7203fa230865a9c42f80809d75c44be17de8feb8b06" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.632998 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-v4rvp" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.635618 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" event={"ID":"5e3bf656-aa9a-4e7c-8d80-78be8d462abb","Type":"ContainerStarted","Data":"86b2911edec33b3ad83ad0811f8c7c1a165b335612c5cddb5ed738e8ce2fe427"} Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.895724 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.961108 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access\") pod \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.961185 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir\") pod \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\" (UID: \"1c7b55b5-32c4-404a-a3f4-e56a6622e581\") " Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.961334 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1c7b55b5-32c4-404a-a3f4-e56a6622e581" (UID: "1c7b55b5-32c4-404a-a3f4-e56a6622e581"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.961651 4836 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:29 crc kubenswrapper[4836]: I0122 09:45:29.969100 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1c7b55b5-32c4-404a-a3f4-e56a6622e581" (UID: "1c7b55b5-32c4-404a-a3f4-e56a6622e581"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:30 crc kubenswrapper[4836]: I0122 09:45:30.064674 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1c7b55b5-32c4-404a-a3f4-e56a6622e581-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:30 crc kubenswrapper[4836]: I0122 09:45:30.649605 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"1c7b55b5-32c4-404a-a3f4-e56a6622e581","Type":"ContainerDied","Data":"af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab"} Jan 22 09:45:30 crc kubenswrapper[4836]: I0122 09:45:30.649649 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 22 09:45:30 crc kubenswrapper[4836]: I0122 09:45:30.649651 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af9019e263c7ad47b4e58a21b6f2ea026d285c9334208e120afc270f73cdc6ab" Jan 22 09:45:31 crc kubenswrapper[4836]: I0122 09:45:31.657650 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" event={"ID":"5e3bf656-aa9a-4e7c-8d80-78be8d462abb","Type":"ContainerStarted","Data":"7708728c85488894516a02b1fbdf27b937f089a46ed33b09aa0dca1a8f1130ec"} Jan 22 09:45:31 crc kubenswrapper[4836]: I0122 09:45:31.659372 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:31 crc kubenswrapper[4836]: I0122 09:45:31.664200 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:31 crc kubenswrapper[4836]: I0122 09:45:31.685133 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" podStartSLOduration=33.685108592 podStartE2EDuration="33.685108592s" podCreationTimestamp="2026-01-22 09:44:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:31.679320547 +0000 UTC m=+215.039089074" watchObservedRunningTime="2026-01-22 09:45:31.685108592 +0000 UTC m=+215.044877119" Jan 22 09:45:34 crc kubenswrapper[4836]: I0122 09:45:34.048587 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:45:56 crc kubenswrapper[4836]: E0122 09:45:56.111321 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 22 09:45:56 crc kubenswrapper[4836]: E0122 09:45:56.112134 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jws8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cpx6h_openshift-marketplace(4363607a-4a5d-4e80-9389-4cd4214e316b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 09:45:56 crc kubenswrapper[4836]: E0122 09:45:56.113442 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cpx6h" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.352124 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.352668 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerName="controller-manager" containerID="cri-o://6764cfdd6a9cc843c1e37806a36d6130c3e7ec7f88acf9415cdee2caec80e519" gracePeriod=30 Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.451092 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.451400 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerName="route-controller-manager" containerID="cri-o://7708728c85488894516a02b1fbdf27b937f089a46ed33b09aa0dca1a8f1130ec" gracePeriod=30 Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.821100 4836 generic.go:334] "Generic (PLEG): container finished" podID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerID="6764cfdd6a9cc843c1e37806a36d6130c3e7ec7f88acf9415cdee2caec80e519" exitCode=0 Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.821169 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" event={"ID":"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656","Type":"ContainerDied","Data":"6764cfdd6a9cc843c1e37806a36d6130c3e7ec7f88acf9415cdee2caec80e519"} Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.822521 4836 generic.go:334] "Generic (PLEG): container finished" podID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerID="7708728c85488894516a02b1fbdf27b937f089a46ed33b09aa0dca1a8f1130ec" exitCode=0 Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.822545 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" event={"ID":"5e3bf656-aa9a-4e7c-8d80-78be8d462abb","Type":"ContainerDied","Data":"7708728c85488894516a02b1fbdf27b937f089a46ed33b09aa0dca1a8f1130ec"} Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.925069 4836 patch_prober.go:28] interesting pod/route-controller-manager-6995bc88c8-4nq46 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Jan 22 09:45:58 crc kubenswrapper[4836]: I0122 09:45:58.925132 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.079542 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" containerName="oauth-openshift" containerID="cri-o://b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6" gracePeriod=15 Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.347631 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.372081 4836 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cbc633b_ae56_438d_88c3_91437b545f38.slice/crio-conmon-b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6.scope\": RecentStats: unable to find data in memory cache]" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.485853 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx8wj\" (UniqueName: \"kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj\") pod \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.485913 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca\") pod \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.485939 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert\") pod \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.485973 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles\") pod \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.486031 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config\") pod \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\" (UID: \"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.487207 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca" (OuterVolumeSpecName: "client-ca") pod "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" (UID: "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.487714 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config" (OuterVolumeSpecName: "config") pod "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" (UID: "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.487906 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" (UID: "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.495308 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" (UID: "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.500952 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.502434 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj" (OuterVolumeSpecName: "kube-api-access-rx8wj") pod "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" (UID: "0fe6ecf9-c1fe-4cb3-abae-2d3a78636656"). InnerVolumeSpecName "kube-api-access-rx8wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.578156 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.586813 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9d5j5\" (UniqueName: \"kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5\") pod \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.586894 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca\") pod \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.586917 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert\") pod \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.586992 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config\") pod \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\" (UID: \"5e3bf656-aa9a-4e7c-8d80-78be8d462abb\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587206 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587225 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587237 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx8wj\" (UniqueName: \"kubernetes.io/projected/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-kube-api-access-rx8wj\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587262 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587270 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.587924 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config" (OuterVolumeSpecName: "config") pod "5e3bf656-aa9a-4e7c-8d80-78be8d462abb" (UID: "5e3bf656-aa9a-4e7c-8d80-78be8d462abb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.588020 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca" (OuterVolumeSpecName: "client-ca") pod "5e3bf656-aa9a-4e7c-8d80-78be8d462abb" (UID: "5e3bf656-aa9a-4e7c-8d80-78be8d462abb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.590941 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5" (OuterVolumeSpecName: "kube-api-access-9d5j5") pod "5e3bf656-aa9a-4e7c-8d80-78be8d462abb" (UID: "5e3bf656-aa9a-4e7c-8d80-78be8d462abb"). InnerVolumeSpecName "kube-api-access-9d5j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.591016 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5e3bf656-aa9a-4e7c-8d80-78be8d462abb" (UID: "5e3bf656-aa9a-4e7c-8d80-78be8d462abb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688584 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688632 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688653 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688688 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688707 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688736 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688757 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688804 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688842 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688860 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688894 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688920 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57ck8\" (UniqueName: \"kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688943 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.688961 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs\") pod \"4cbc633b-ae56-438d-88c3-91437b545f38\" (UID: \"4cbc633b-ae56-438d-88c3-91437b545f38\") " Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.689181 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.689192 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9d5j5\" (UniqueName: \"kubernetes.io/projected/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-kube-api-access-9d5j5\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.689201 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.689209 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e3bf656-aa9a-4e7c-8d80-78be8d462abb-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.689682 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.690287 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.690394 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.690403 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.691101 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.695909 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8" (OuterVolumeSpecName: "kube-api-access-57ck8") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "kube-api-access-57ck8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.700730 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.702720 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.703959 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.704395 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.705269 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.705629 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.713561 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.715248 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4cbc633b-ae56-438d-88c3-91437b545f38" (UID: "4cbc633b-ae56-438d-88c3-91437b545f38"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790502 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790860 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790877 4836 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4cbc633b-ae56-438d-88c3-91437b545f38-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790895 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790910 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57ck8\" (UniqueName: \"kubernetes.io/projected/4cbc633b-ae56-438d-88c3-91437b545f38-kube-api-access-57ck8\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790922 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790937 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790949 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790962 4836 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790975 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.790987 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.791000 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.791012 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.791025 4836 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4cbc633b-ae56-438d-88c3-91437b545f38-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.829598 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerStarted","Data":"cab067f8b7d4b0b584700bf162084895d67a82366e85cd69d45665c958f885d0"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.831132 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerStarted","Data":"165a17f00de4a0ed6b515b23e96577d2155575321e967ebaec84c5617f85ea46"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.832372 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" event={"ID":"5e3bf656-aa9a-4e7c-8d80-78be8d462abb","Type":"ContainerDied","Data":"86b2911edec33b3ad83ad0811f8c7c1a165b335612c5cddb5ed738e8ce2fe427"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.832395 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.832476 4836 scope.go:117] "RemoveContainer" containerID="7708728c85488894516a02b1fbdf27b937f089a46ed33b09aa0dca1a8f1130ec" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.852740 4836 generic.go:334] "Generic (PLEG): container finished" podID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerID="66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6" exitCode=0 Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.852817 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerDied","Data":"66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.860336 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.860360 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b" event={"ID":"0fe6ecf9-c1fe-4cb3-abae-2d3a78636656","Type":"ContainerDied","Data":"3c82540160b00998f6b4c0fd7c100efdbc40d6c935e6dfcfdceda02a16f535f9"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.868935 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerStarted","Data":"a2644ef719bbefec72d05031434fac6410e5e6f6f0f0241229bd289c03a41053"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.872981 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerStarted","Data":"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.874317 4836 generic.go:334] "Generic (PLEG): container finished" podID="4cbc633b-ae56-438d-88c3-91437b545f38" containerID="b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6" exitCode=0 Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.874351 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.874372 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" event={"ID":"4cbc633b-ae56-438d-88c3-91437b545f38","Type":"ContainerDied","Data":"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.874436 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-jwkz2" event={"ID":"4cbc633b-ae56-438d-88c3-91437b545f38","Type":"ContainerDied","Data":"d611aca7aa9f6a4396c0976f15161ea4233162f9d4ab35a527641b65cba09e0e"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.879754 4836 generic.go:334] "Generic (PLEG): container finished" podID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerID="acb68c99287e899201287e5437914995784480f2adc37e70d8d4caa5abddeb1a" exitCode=0 Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.879831 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerDied","Data":"acb68c99287e899201287e5437914995784480f2adc37e70d8d4caa5abddeb1a"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.884835 4836 generic.go:334] "Generic (PLEG): container finished" podID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerID="da71b4d14abfbe08e3081ff13ceb0bc2c9bee41727ad2596a91aec994053679c" exitCode=0 Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.884868 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerDied","Data":"da71b4d14abfbe08e3081ff13ceb0bc2c9bee41727ad2596a91aec994053679c"} Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996184 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.996472 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b43b9c-abd0-4c0f-91d5-d92477b97487" containerName="collect-profiles" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996491 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b43b9c-abd0-4c0f-91d5-d92477b97487" containerName="collect-profiles" Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.996503 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" containerName="oauth-openshift" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996515 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" containerName="oauth-openshift" Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.996528 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c7b55b5-32c4-404a-a3f4-e56a6622e581" containerName="pruner" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996535 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c7b55b5-32c4-404a-a3f4-e56a6622e581" containerName="pruner" Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.996545 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerName="controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996552 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerName="controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: E0122 09:45:59.996560 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerName="route-controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996567 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerName="route-controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996688 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" containerName="oauth-openshift" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996701 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b43b9c-abd0-4c0f-91d5-d92477b97487" containerName="collect-profiles" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996711 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" containerName="route-controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996723 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c7b55b5-32c4-404a-a3f4-e56a6622e581" containerName="pruner" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.996733 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" containerName="controller-manager" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.997200 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.998670 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:45:59 crc kubenswrapper[4836]: I0122 09:45:59.999384 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.000816 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.000942 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.001510 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.001840 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.002007 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.002086 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.002240 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.004533 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.005232 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.005266 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.005235 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.005427 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.007320 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.022382 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.024930 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096321 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096367 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvsz5\" (UniqueName: \"kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096395 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096416 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096443 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d7vn\" (UniqueName: \"kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096461 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096500 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096516 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.096532 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.109144 4836 scope.go:117] "RemoveContainer" containerID="6764cfdd6a9cc843c1e37806a36d6130c3e7ec7f88acf9415cdee2caec80e519" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.134790 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.135706 4836 scope.go:117] "RemoveContainer" containerID="b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.139542 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6995bc88c8-4nq46"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.148952 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.155602 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6ff6dd6f76-llz2b"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.156204 4836 scope.go:117] "RemoveContainer" containerID="b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6" Jan 22 09:46:00 crc kubenswrapper[4836]: E0122 09:46:00.156677 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6\": container with ID starting with b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6 not found: ID does not exist" containerID="b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.156712 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6"} err="failed to get container status \"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6\": rpc error: code = NotFound desc = could not find container \"b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6\": container with ID starting with b75834b67a6d8417dd43534b40a1d2a940d83773fd4b1a2414ce74af3e11a5f6 not found: ID does not exist" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.158705 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.161676 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-jwkz2"] Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197363 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d7vn\" (UniqueName: \"kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197435 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197487 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197510 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197530 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197577 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197605 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvsz5\" (UniqueName: \"kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197629 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.197653 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.198479 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.199055 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.199459 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.199557 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.200745 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.204062 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.213267 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d7vn\" (UniqueName: \"kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.213517 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvsz5\" (UniqueName: \"kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5\") pod \"route-controller-manager-7d9fb8b77c-bdmvc\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.214970 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert\") pod \"controller-manager-d48677986-ttmcf\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.438265 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.448564 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.647283 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:46:00 crc kubenswrapper[4836]: W0122 09:46:00.664775 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17b4ea51_6711_4c06_a17e_5a26b0590fa6.slice/crio-3d054b69866190511ffb2fdbb38bd52c29e09552413568de26daae6a64ad4029 WatchSource:0}: Error finding container 3d054b69866190511ffb2fdbb38bd52c29e09552413568de26daae6a64ad4029: Status 404 returned error can't find the container with id 3d054b69866190511ffb2fdbb38bd52c29e09552413568de26daae6a64ad4029 Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.681302 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:46:00 crc kubenswrapper[4836]: W0122 09:46:00.687898 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod794eb7c1_6182_4792_9798_5553ebf10ad4.slice/crio-faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a WatchSource:0}: Error finding container faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a: Status 404 returned error can't find the container with id faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.893609 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" event={"ID":"794eb7c1-6182-4792-9798-5553ebf10ad4","Type":"ContainerStarted","Data":"bf65d926788eeb3b71e09c48eed24099087fef5838a02fb17c54f914a428d11f"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.893666 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" event={"ID":"794eb7c1-6182-4792-9798-5553ebf10ad4","Type":"ContainerStarted","Data":"faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.893891 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.895644 4836 patch_prober.go:28] interesting pod/route-controller-manager-7d9fb8b77c-bdmvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.895687 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.896200 4836 generic.go:334] "Generic (PLEG): container finished" podID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerID="a2644ef719bbefec72d05031434fac6410e5e6f6f0f0241229bd289c03a41053" exitCode=0 Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.896264 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerDied","Data":"a2644ef719bbefec72d05031434fac6410e5e6f6f0f0241229bd289c03a41053"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.899106 4836 generic.go:334] "Generic (PLEG): container finished" podID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerID="cab067f8b7d4b0b584700bf162084895d67a82366e85cd69d45665c958f885d0" exitCode=0 Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.899148 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerDied","Data":"cab067f8b7d4b0b584700bf162084895d67a82366e85cd69d45665c958f885d0"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.902338 4836 generic.go:334] "Generic (PLEG): container finished" podID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerID="165a17f00de4a0ed6b515b23e96577d2155575321e967ebaec84c5617f85ea46" exitCode=0 Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.902402 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerDied","Data":"165a17f00de4a0ed6b515b23e96577d2155575321e967ebaec84c5617f85ea46"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.909057 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerStarted","Data":"7414f3234b7b8f2b7cf7c340e91226d1f57f742820b1eddb0c7d52fe22d2942f"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.910530 4836 generic.go:334] "Generic (PLEG): container finished" podID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerID="80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6" exitCode=0 Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.910580 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerDied","Data":"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.915435 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerStarted","Data":"9e2f2bfeec75c92b16687fdafb4ea62255a7150d193bcd852a5117c2990ad173"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.917343 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" podStartSLOduration=2.917332899 podStartE2EDuration="2.917332899s" podCreationTimestamp="2026-01-22 09:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:00.915389287 +0000 UTC m=+244.275157824" watchObservedRunningTime="2026-01-22 09:46:00.917332899 +0000 UTC m=+244.277101416" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.920777 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" event={"ID":"17b4ea51-6711-4c06-a17e-5a26b0590fa6","Type":"ContainerStarted","Data":"a838fbcfa417f78c2b3a1fd03ad4732494737cdb79000727629fe38bf25c1d1c"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.920808 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" event={"ID":"17b4ea51-6711-4c06-a17e-5a26b0590fa6","Type":"ContainerStarted","Data":"3d054b69866190511ffb2fdbb38bd52c29e09552413568de26daae6a64ad4029"} Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.921393 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.927586 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:46:00 crc kubenswrapper[4836]: I0122 09:46:00.934955 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ztdcc" podStartSLOduration=3.684347198 podStartE2EDuration="1m33.93493656s" podCreationTimestamp="2026-01-22 09:44:27 +0000 UTC" firstStartedPulling="2026-01-22 09:44:30.207410522 +0000 UTC m=+153.567179039" lastFinishedPulling="2026-01-22 09:46:00.457999884 +0000 UTC m=+243.817768401" observedRunningTime="2026-01-22 09:46:00.933528543 +0000 UTC m=+244.293297060" watchObservedRunningTime="2026-01-22 09:46:00.93493656 +0000 UTC m=+244.294705077" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.028794 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fe6ecf9-c1fe-4cb3-abae-2d3a78636656" path="/var/lib/kubelet/pods/0fe6ecf9-c1fe-4cb3-abae-2d3a78636656/volumes" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.030894 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cbc633b-ae56-438d-88c3-91437b545f38" path="/var/lib/kubelet/pods/4cbc633b-ae56-438d-88c3-91437b545f38/volumes" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.031868 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e3bf656-aa9a-4e7c-8d80-78be8d462abb" path="/var/lib/kubelet/pods/5e3bf656-aa9a-4e7c-8d80-78be8d462abb/volumes" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.059488 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" podStartSLOduration=3.059464923 podStartE2EDuration="3.059464923s" podCreationTimestamp="2026-01-22 09:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:01.057777888 +0000 UTC m=+244.417546405" watchObservedRunningTime="2026-01-22 09:46:01.059464923 +0000 UTC m=+244.419233440" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.087594 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-86vhq" podStartSLOduration=3.670723502 podStartE2EDuration="1m33.087575926s" podCreationTimestamp="2026-01-22 09:44:28 +0000 UTC" firstStartedPulling="2026-01-22 09:44:31.183630184 +0000 UTC m=+154.543398701" lastFinishedPulling="2026-01-22 09:46:00.600482608 +0000 UTC m=+243.960251125" observedRunningTime="2026-01-22 09:46:01.084886564 +0000 UTC m=+244.444655081" watchObservedRunningTime="2026-01-22 09:46:01.087575926 +0000 UTC m=+244.447344443" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.929849 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerStarted","Data":"ef23fc6e1e30992548a6a333e18e8d1c11bed5e79ee0d6524a9ed3610e2d6954"} Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.932104 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerStarted","Data":"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29"} Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.936343 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerStarted","Data":"36770dd1b73f0c1834bd32c08cc61f90a2e026f707c90fe90850d12162500a94"} Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.939411 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerStarted","Data":"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c"} Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.944133 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.953524 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2xnzx" podStartSLOduration=9.681935784 podStartE2EDuration="1m32.953494974s" podCreationTimestamp="2026-01-22 09:44:29 +0000 UTC" firstStartedPulling="2026-01-22 09:44:38.273861391 +0000 UTC m=+161.633629908" lastFinishedPulling="2026-01-22 09:46:01.545420581 +0000 UTC m=+244.905189098" observedRunningTime="2026-01-22 09:46:01.950643448 +0000 UTC m=+245.310411955" watchObservedRunningTime="2026-01-22 09:46:01.953494974 +0000 UTC m=+245.313263491" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.991527 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zbvh5" podStartSLOduration=2.396357399 podStartE2EDuration="1m35.991508711s" podCreationTimestamp="2026-01-22 09:44:26 +0000 UTC" firstStartedPulling="2026-01-22 09:44:28.111599528 +0000 UTC m=+151.471368035" lastFinishedPulling="2026-01-22 09:46:01.70675083 +0000 UTC m=+245.066519347" observedRunningTime="2026-01-22 09:46:01.988703116 +0000 UTC m=+245.348471643" watchObservedRunningTime="2026-01-22 09:46:01.991508711 +0000 UTC m=+245.351277228" Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.998236 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-788784fd4b-nn4b6"] Jan 22 09:46:01 crc kubenswrapper[4836]: I0122 09:46:01.998872 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.003744 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.003865 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.003897 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004089 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004175 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004241 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004523 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004646 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.004921 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.005702 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.005915 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.014108 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.021006 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.026878 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.053652 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.059353 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-788784fd4b-nn4b6"] Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.124701 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n4dwm" podStartSLOduration=3.40959699 podStartE2EDuration="1m36.124679197s" podCreationTimestamp="2026-01-22 09:44:26 +0000 UTC" firstStartedPulling="2026-01-22 09:44:28.111274237 +0000 UTC m=+151.471042754" lastFinishedPulling="2026-01-22 09:46:00.826356444 +0000 UTC m=+244.186124961" observedRunningTime="2026-01-22 09:46:02.085870698 +0000 UTC m=+245.445639215" watchObservedRunningTime="2026-01-22 09:46:02.124679197 +0000 UTC m=+245.484447714" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.125067 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7bfkp" podStartSLOduration=3.637725091 podStartE2EDuration="1m37.125060677s" podCreationTimestamp="2026-01-22 09:44:25 +0000 UTC" firstStartedPulling="2026-01-22 09:44:28.150478298 +0000 UTC m=+151.510246815" lastFinishedPulling="2026-01-22 09:46:01.637813884 +0000 UTC m=+244.997582401" observedRunningTime="2026-01-22 09:46:02.122171589 +0000 UTC m=+245.481940106" watchObservedRunningTime="2026-01-22 09:46:02.125060677 +0000 UTC m=+245.484829214" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.128294 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-service-ca\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129137 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddb2r\" (UniqueName: \"kubernetes.io/projected/968371c2-dd96-4673-b44e-2ebff3e93a44-kube-api-access-ddb2r\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129168 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-login\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129317 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129817 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-error\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129840 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-policies\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129940 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-cliconfig\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.129993 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-router-certs\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130063 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130100 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-serving-cert\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130169 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-session\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130244 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130303 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.130337 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-dir\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232473 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-cliconfig\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232559 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-router-certs\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232586 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232611 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-serving-cert\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232645 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-session\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232672 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232699 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.232718 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-dir\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.233426 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-cliconfig\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234149 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234226 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-service-ca\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234260 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-dir\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234273 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddb2r\" (UniqueName: \"kubernetes.io/projected/968371c2-dd96-4673-b44e-2ebff3e93a44-kube-api-access-ddb2r\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234296 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-login\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234339 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234360 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-error\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234382 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-policies\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.234877 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-audit-policies\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.235009 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-service-ca\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.240656 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-error\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.240857 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.241468 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-router-certs\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.242665 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-session\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.242811 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-login\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.245203 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.248746 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.249194 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/968371c2-dd96-4673-b44e-2ebff3e93a44-v4-0-config-system-serving-cert\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.259954 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddb2r\" (UniqueName: \"kubernetes.io/projected/968371c2-dd96-4673-b44e-2ebff3e93a44-kube-api-access-ddb2r\") pod \"oauth-openshift-788784fd4b-nn4b6\" (UID: \"968371c2-dd96-4673-b44e-2ebff3e93a44\") " pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.317643 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.762004 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-788784fd4b-nn4b6"] Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.949904 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" event={"ID":"968371c2-dd96-4673-b44e-2ebff3e93a44","Type":"ContainerStarted","Data":"de70eb0c8d4d726aa057e9c11dc26744bdcbb70f6d3c409dc0bd658dab79963c"} Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.953056 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerStarted","Data":"fb93e0e0be56944808e5cd76ef6d710fbff5cd8eddfca195dc26a0ed9e999695"} Jan 22 09:46:02 crc kubenswrapper[4836]: I0122 09:46:02.983534 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6pq7n" podStartSLOduration=4.279535008 podStartE2EDuration="1m37.983485166s" podCreationTimestamp="2026-01-22 09:44:25 +0000 UTC" firstStartedPulling="2026-01-22 09:44:28.128489866 +0000 UTC m=+151.488258373" lastFinishedPulling="2026-01-22 09:46:01.832440014 +0000 UTC m=+245.192208531" observedRunningTime="2026-01-22 09:46:02.980402332 +0000 UTC m=+246.340170859" watchObservedRunningTime="2026-01-22 09:46:02.983485166 +0000 UTC m=+246.343253703" Jan 22 09:46:03 crc kubenswrapper[4836]: I0122 09:46:03.960204 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" event={"ID":"968371c2-dd96-4673-b44e-2ebff3e93a44","Type":"ContainerStarted","Data":"309bd38af1f3595aa183b2528f696ffe04047494169680c0258908bf143f5902"} Jan 22 09:46:03 crc kubenswrapper[4836]: I0122 09:46:03.961021 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:03 crc kubenswrapper[4836]: I0122 09:46:03.968280 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" Jan 22 09:46:03 crc kubenswrapper[4836]: I0122 09:46:03.983344 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-788784fd4b-nn4b6" podStartSLOduration=29.983327778 podStartE2EDuration="29.983327778s" podCreationTimestamp="2026-01-22 09:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:03.983078662 +0000 UTC m=+247.342847179" watchObservedRunningTime="2026-01-22 09:46:03.983327778 +0000 UTC m=+247.343096295" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.419543 4836 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.420659 4836 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.420765 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.420973 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6" gracePeriod=15 Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.421016 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10" gracePeriod=15 Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.421116 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e" gracePeriod=15 Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.421120 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22" gracePeriod=15 Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.421129 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d" gracePeriod=15 Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423127 4836 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423295 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423312 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423322 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423328 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423335 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423342 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423352 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423357 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423364 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423370 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423379 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423385 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: E0122 09:46:05.423394 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423399 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423493 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423502 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423510 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423517 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423524 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.423537 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.465858 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.481988 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.482332 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.482838 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.482922 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.483014 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.483090 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.483132 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.483157 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584483 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584555 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584619 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584640 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584665 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584733 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584742 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584772 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584838 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584889 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584913 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584902 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584929 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584958 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.584985 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.585107 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.763136 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:46:05 crc kubenswrapper[4836]: W0122 09:46:05.791615 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-33ae9702355ebb996fdb103ebc5f4fcf52021e02d8ddb694dd6a2486525651cd WatchSource:0}: Error finding container 33ae9702355ebb996fdb103ebc5f4fcf52021e02d8ddb694dd6a2486525651cd: Status 404 returned error can't find the container with id 33ae9702355ebb996fdb103ebc5f4fcf52021e02d8ddb694dd6a2486525651cd Jan 22 09:46:05 crc kubenswrapper[4836]: I0122 09:46:05.973556 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"33ae9702355ebb996fdb103ebc5f4fcf52021e02d8ddb694dd6a2486525651cd"} Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.121549 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.122114 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.306076 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.306125 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.500899 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.500954 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.611998 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.612087 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.612987 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.613436 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.613966 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.614509 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.614962 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.615512 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.616002 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.616429 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.616907 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.617330 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.759499 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.759566 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.804775 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.805391 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.805834 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.806154 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.806402 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.806653 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.806896 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.980912 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.982191 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.982728 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e" exitCode=0 Jan 22 09:46:06 crc kubenswrapper[4836]: I0122 09:46:06.982749 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22" exitCode=2 Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.002310 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.003707 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.003977 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.004292 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.004545 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.004791 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.028170 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.028707 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.028954 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.029135 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.029422 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.029661 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.029889 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.030168 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.030395 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.030617 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.030859 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.031105 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.031719 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.032002 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.032282 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.032518 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.032767 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.032966 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.037369 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.037662 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.037981 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.038311 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.038569 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.038790 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: E0122 09:46:07.337382 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d047de213fcfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,LastTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:46:07 crc kubenswrapper[4836]: E0122 09:46:07.950400 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d047de213fcfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,LastTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.990876 4836 generic.go:334] "Generic (PLEG): container finished" podID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" containerID="d491c6962735d4108d8e22f896d7185b9f600798596cf71691ab6418042bf9f8" exitCode=0 Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.990946 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63b6b777-0f50-4dd2-af16-1f27e8969a7f","Type":"ContainerDied","Data":"d491c6962735d4108d8e22f896d7185b9f600798596cf71691ab6418042bf9f8"} Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.992031 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.993251 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.993624 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.994133 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.994454 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.994708 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.995256 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2"} Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.998108 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 22 09:46:07 crc kubenswrapper[4836]: I0122 09:46:07.999850 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.001269 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10" exitCode=0 Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.001561 4836 scope.go:117] "RemoveContainer" containerID="ec8bbe0cdf4851309620ee0ff78d98f0e075b47f5d42d037d8036cd8a2407fb7" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.106423 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.106787 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.168714 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.169439 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.169698 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.169925 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.171601 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.175543 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.175871 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.177979 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.524751 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.525208 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.560858 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.561461 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.561800 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.561970 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.562207 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.562439 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.562739 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.563212 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:08 crc kubenswrapper[4836]: I0122 09:46:08.563431 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.009824 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.011083 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d" exitCode=0 Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.011112 4836 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6" exitCode=0 Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.013174 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.013577 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.013773 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.013929 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.014371 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.014570 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.014745 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.015029 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.053640 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.054166 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.055316 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.055575 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.055779 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.056084 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.056806 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.057116 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.057382 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.062701 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.064917 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.065268 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.065474 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.065615 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.065745 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.065940 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.066277 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.066577 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.330110 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.330896 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.331309 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.331756 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.332089 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.332377 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.332671 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.332963 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.333252 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.383690 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.384844 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.385323 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.385644 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.385990 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.386298 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.386553 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.386956 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.387228 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.387531 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.387921 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438352 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock\") pod \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438440 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock" (OuterVolumeSpecName: "var-lock") pod "63b6b777-0f50-4dd2-af16-1f27e8969a7f" (UID: "63b6b777-0f50-4dd2-af16-1f27e8969a7f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438462 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438521 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access\") pod \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438532 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438548 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir\") pod \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\" (UID: \"63b6b777-0f50-4dd2-af16-1f27e8969a7f\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438572 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438603 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438628 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438629 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "63b6b777-0f50-4dd2-af16-1f27e8969a7f" (UID: "63b6b777-0f50-4dd2-af16-1f27e8969a7f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438727 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.438998 4836 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.439010 4836 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.439020 4836 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.439028 4836 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.439052 4836 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.457343 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "63b6b777-0f50-4dd2-af16-1f27e8969a7f" (UID: "63b6b777-0f50-4dd2-af16-1f27e8969a7f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.539784 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63b6b777-0f50-4dd2-af16-1f27e8969a7f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.908776 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.909136 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.955496 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.956425 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.956891 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.957172 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.957425 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.957683 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.957928 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.958193 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.958459 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.958698 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:09 crc kubenswrapper[4836]: I0122 09:46:09.958940 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.018391 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.019029 4836 scope.go:117] "RemoveContainer" containerID="26c83787907868d9cb61d0f5d3e303e128c1c29183bf242c7e9aad976976bc10" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.019176 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.033802 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.037861 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.038285 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.038825 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.039327 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.039822 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.040359 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.040900 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.041455 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.041953 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.042487 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.044482 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63b6b777-0f50-4dd2-af16-1f27e8969a7f","Type":"ContainerDied","Data":"e9e08e92b548f6c42cdb64fb44e9c2e02381457938959655a3c78ef739e99d80"} Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.044523 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e08e92b548f6c42cdb64fb44e9c2e02381457938959655a3c78ef739e99d80" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.052484 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.053618 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.053840 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.054066 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.054265 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.054456 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.054720 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.054924 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.055139 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.055333 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.055896 4836 scope.go:117] "RemoveContainer" containerID="094c4d7647a91a9664f1ff747e6ef18961fb6151a2ef268b451ad141d3483b4e" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.072366 4836 scope.go:117] "RemoveContainer" containerID="7f5f688710c8e15fa51e9ab02695d6e692141f8eb34c66682e1a8af350a86c2d" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.104000 4836 scope.go:117] "RemoveContainer" containerID="020fb7c6d64a3d02f61a8a0226bed40fb66475297efb51b5d22ead5f350aaf22" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.104622 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.105146 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.105358 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.105553 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.105780 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.106197 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.106439 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.106664 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.106902 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.107137 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.107363 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.123263 4836 scope.go:117] "RemoveContainer" containerID="96b829bdcd1212509d7281245c864d680c29d587759ae915f40e806d30cdfcd6" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.138477 4836 scope.go:117] "RemoveContainer" containerID="60d92ad9f1bc4324f0a86fd6523dd06fff5fc9fc669d18456404fd25a57da88c" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.690880 4836 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.691467 4836 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.692110 4836 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.692647 4836 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.693238 4836 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.693318 4836 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.693744 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="200ms" Jan 22 09:46:10 crc kubenswrapper[4836]: E0122 09:46:10.894606 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="400ms" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.997461 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.997956 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.998327 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.998871 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.999250 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.999489 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:10 crc kubenswrapper[4836]: I0122 09:46:10.999779 4836 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:11 crc kubenswrapper[4836]: I0122 09:46:11.000137 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:11 crc kubenswrapper[4836]: I0122 09:46:11.000496 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:11 crc kubenswrapper[4836]: I0122 09:46:11.000819 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:11 crc kubenswrapper[4836]: E0122 09:46:11.001000 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cpx6h" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" Jan 22 09:46:11 crc kubenswrapper[4836]: I0122 09:46:11.001099 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:11 crc kubenswrapper[4836]: I0122 09:46:11.003862 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 22 09:46:11 crc kubenswrapper[4836]: E0122 09:46:11.296081 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="800ms" Jan 22 09:46:12 crc kubenswrapper[4836]: E0122 09:46:12.096735 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="1.6s" Jan 22 09:46:13 crc kubenswrapper[4836]: E0122 09:46:13.698124 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="3.2s" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.757726 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:46:14Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:46:14Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:46:14Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:46:14Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.758377 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.758926 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.759621 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.760168 4836 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:14 crc kubenswrapper[4836]: E0122 09:46:14.760220 4836 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 22 09:46:16 crc kubenswrapper[4836]: E0122 09:46:16.898945 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="6.4s" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.001264 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.001849 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.002387 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.002782 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.003402 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.003854 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.004299 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.004771 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.005237 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.005645 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:17 crc kubenswrapper[4836]: I0122 09:46:17.376277 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-6pq7n" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="registry-server" probeResult="failure" output=< Jan 22 09:46:17 crc kubenswrapper[4836]: timeout: failed to connect service ":50051" within 1s Jan 22 09:46:17 crc kubenswrapper[4836]: > Jan 22 09:46:17 crc kubenswrapper[4836]: E0122 09:46:17.951787 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d047de213fcfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,LastTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:46:18 crc kubenswrapper[4836]: I0122 09:46:18.996315 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:18 crc kubenswrapper[4836]: I0122 09:46:18.998093 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:18 crc kubenswrapper[4836]: I0122 09:46:18.998596 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:18 crc kubenswrapper[4836]: I0122 09:46:18.999191 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:18.999939 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.000435 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.000793 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.001283 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.001722 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.002072 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.002443 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.019973 4836 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.020013 4836 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:19 crc kubenswrapper[4836]: E0122 09:46:19.021145 4836 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.022091 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:19 crc kubenswrapper[4836]: W0122 09:46:19.061382 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-9b59d11f38d9ab4a1f90649967aa6775291941424d0d2cac202b5316246e5e29 WatchSource:0}: Error finding container 9b59d11f38d9ab4a1f90649967aa6775291941424d0d2cac202b5316246e5e29: Status 404 returned error can't find the container with id 9b59d11f38d9ab4a1f90649967aa6775291941424d0d2cac202b5316246e5e29 Jan 22 09:46:19 crc kubenswrapper[4836]: I0122 09:46:19.098191 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9b59d11f38d9ab4a1f90649967aa6775291941424d0d2cac202b5316246e5e29"} Jan 22 09:46:23 crc kubenswrapper[4836]: I0122 09:46:23.020580 4836 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 22 09:46:23 crc kubenswrapper[4836]: I0122 09:46:23.021348 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 22 09:46:23 crc kubenswrapper[4836]: E0122 09:46:23.301254 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="7s" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.007875 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.008499 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.009905 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.010699 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.011367 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.011898 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.012460 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.012981 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.013389 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:24 crc kubenswrapper[4836]: I0122 09:46:24.013918 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:25 crc kubenswrapper[4836]: I0122 09:46:25.142463 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7fdba623f31dfc6c42d2eebf64271a27f5e0970c5c1ead993b72e727e59ce88e"} Jan 22 09:46:25 crc kubenswrapper[4836]: I0122 09:46:25.822837 4836 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 22 09:46:25 crc kubenswrapper[4836]: I0122 09:46:25.823181 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.007315 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.008146 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.008696 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.009145 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.009361 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.009550 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.009724 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.009905 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.010126 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.010307 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: I0122 09:46:27.010531 4836 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:27 crc kubenswrapper[4836]: E0122 09:46:27.953458 4836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.173:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d047de213fcfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,LastTimestamp:2026-01-22 09:46:07.336856826 +0000 UTC m=+250.696625343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.176738 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.177216 4836 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18" exitCode=1 Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.177330 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18"} Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.178203 4836 scope.go:117] "RemoveContainer" containerID="bb20c32aafc8581b559a977a3e0fc8b1c3895ce5be809d9e52aab9c8c7ce2f18" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.178340 4836 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.178653 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.179191 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.179762 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.180063 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.180439 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.181118 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.181334 4836 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="7fdba623f31dfc6c42d2eebf64271a27f5e0970c5c1ead993b72e727e59ce88e" exitCode=0 Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.181393 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"7fdba623f31dfc6c42d2eebf64271a27f5e0970c5c1ead993b72e727e59ce88e"} Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.181628 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.182158 4836 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.182203 4836 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.182265 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.182706 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: E0122 09:46:29.182741 4836 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.182938 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.183146 4836 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.183445 4836 status_manager.go:851] "Failed to get status for pod" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" pod="openshift-marketplace/redhat-operators-cpx6h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-cpx6h\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.183754 4836 status_manager.go:851] "Failed to get status for pod" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" pod="openshift-marketplace/redhat-operators-2xnzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-2xnzx\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.184272 4836 status_manager.go:851] "Failed to get status for pod" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" pod="openshift-marketplace/redhat-marketplace-ztdcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ztdcc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.184723 4836 status_manager.go:851] "Failed to get status for pod" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" pod="openshift-marketplace/community-operators-7bfkp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-7bfkp\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.185210 4836 status_manager.go:851] "Failed to get status for pod" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" pod="openshift-marketplace/redhat-marketplace-86vhq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-86vhq\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.185435 4836 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.185782 4836 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.186292 4836 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.186808 4836 status_manager.go:851] "Failed to get status for pod" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" pod="openshift-marketplace/community-operators-zbvh5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-zbvh5\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.187326 4836 status_manager.go:851] "Failed to get status for pod" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" pod="openshift-marketplace/certified-operators-6pq7n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6pq7n\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.187768 4836 status_manager.go:851] "Failed to get status for pod" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" pod="openshift-marketplace/certified-operators-n4dwm" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-n4dwm\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:29 crc kubenswrapper[4836]: I0122 09:46:29.188272 4836 status_manager.go:851] "Failed to get status for pod" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.173:6443: connect: connection refused" Jan 22 09:46:30 crc kubenswrapper[4836]: E0122 09:46:30.302701 4836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.173:6443: connect: connection refused" interval="7s" Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.195258 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.195784 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e97671675d095c2b4a26994252f4debccc526b50148d40df178f4a67e405e278"} Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.199016 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c11b32b95cd0fd3e698c9ec9379d0466b80b093a744466c2a031f9c93ca23357"} Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.199186 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2702e3d60a3ccc57ef4a8fe94cbb96b9e653ee409922ecc47c266a95efdadf71"} Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.938403 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:46:31 crc kubenswrapper[4836]: I0122 09:46:31.942878 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:46:32 crc kubenswrapper[4836]: I0122 09:46:32.206219 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:46:33 crc kubenswrapper[4836]: I0122 09:46:33.216554 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerStarted","Data":"83bbbbbfea658121a9494c11abdda321e1e6a6af25b144e68c4dc6c32e2f327b"} Jan 22 09:46:33 crc kubenswrapper[4836]: I0122 09:46:33.220118 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ae79812d19b8a36b0e202ae65d881da9b1afda499989c6a4e8b8652371dfe491"} Jan 22 09:46:35 crc kubenswrapper[4836]: I0122 09:46:35.237737 4836 generic.go:334] "Generic (PLEG): container finished" podID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerID="83bbbbbfea658121a9494c11abdda321e1e6a6af25b144e68c4dc6c32e2f327b" exitCode=0 Jan 22 09:46:35 crc kubenswrapper[4836]: I0122 09:46:35.237861 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerDied","Data":"83bbbbbfea658121a9494c11abdda321e1e6a6af25b144e68c4dc6c32e2f327b"} Jan 22 09:46:35 crc kubenswrapper[4836]: I0122 09:46:35.243219 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"40a09e579d763f7d10551f64cc603bdfb3c0a4647a61b4e7eab9bdc130dca237"} Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.261651 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7274bd6e4b3f8c2644f247127a1ae051b6e73e2e5df8f8ac38a28f44e65b8720"} Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.262010 4836 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.262033 4836 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.262171 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.271216 4836 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:46:36 crc kubenswrapper[4836]: I0122 09:46:36.753669 4836 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c82ec481-74a8-4eed-8eb4-4f133bde6585" Jan 22 09:46:37 crc kubenswrapper[4836]: I0122 09:46:37.268496 4836 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:37 crc kubenswrapper[4836]: I0122 09:46:37.268539 4836 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8da4f34-5379-4b90-8c5b-e33691b4107e" Jan 22 09:46:37 crc kubenswrapper[4836]: I0122 09:46:37.273131 4836 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c82ec481-74a8-4eed-8eb4-4f133bde6585" Jan 22 09:46:38 crc kubenswrapper[4836]: I0122 09:46:38.276370 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerStarted","Data":"48c3d6f1bc3a105595acf43ce9dd9a638cda668ec974014d73694c7cc3a0db1a"} Jan 22 09:46:39 crc kubenswrapper[4836]: I0122 09:46:39.527452 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:46:39 crc kubenswrapper[4836]: I0122 09:46:39.527705 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:46:40 crc kubenswrapper[4836]: I0122 09:46:40.580703 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cpx6h" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="registry-server" probeResult="failure" output=< Jan 22 09:46:40 crc kubenswrapper[4836]: timeout: failed to connect service ":50051" within 1s Jan 22 09:46:40 crc kubenswrapper[4836]: > Jan 22 09:46:43 crc kubenswrapper[4836]: I0122 09:46:43.024639 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 22 09:46:49 crc kubenswrapper[4836]: I0122 09:46:49.593739 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:46:49 crc kubenswrapper[4836]: I0122 09:46:49.630407 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 22 09:46:49 crc kubenswrapper[4836]: I0122 09:46:49.638891 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:46:56 crc kubenswrapper[4836]: I0122 09:46:56.825526 4836 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 22 09:47:01 crc kubenswrapper[4836]: I0122 09:47:01.278779 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 22 09:47:03 crc kubenswrapper[4836]: I0122 09:47:03.536634 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 22 09:47:04 crc kubenswrapper[4836]: I0122 09:47:04.125590 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 22 09:47:05 crc kubenswrapper[4836]: I0122 09:47:05.584291 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 22 09:47:06 crc kubenswrapper[4836]: I0122 09:47:06.405729 4836 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 22 09:47:08 crc kubenswrapper[4836]: I0122 09:47:08.235258 4836 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 22 09:47:08 crc kubenswrapper[4836]: I0122 09:47:08.307224 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 22 09:47:08 crc kubenswrapper[4836]: I0122 09:47:08.338011 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 22 09:47:08 crc kubenswrapper[4836]: I0122 09:47:08.986836 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.109335 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.236257 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.473958 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.498953 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.566284 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 22 09:47:09 crc kubenswrapper[4836]: I0122 09:47:09.676292 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.055891 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.075782 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.137396 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.185636 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.201123 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 22 09:47:10 crc kubenswrapper[4836]: I0122 09:47:10.903315 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.062863 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.164899 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.166443 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.442639 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.553157 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.731381 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 22 09:47:11 crc kubenswrapper[4836]: I0122 09:47:11.923602 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.243805 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.311434 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.348183 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.365588 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.673262 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 22 09:47:12 crc kubenswrapper[4836]: I0122 09:47:12.992130 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.379445 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.480325 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.514671 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.616023 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.771816 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:47:13 crc kubenswrapper[4836]: I0122 09:47:13.773710 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.084114 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.123178 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.166805 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.198104 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.254287 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.696188 4836 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.705835 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.888836 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.903424 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 22 09:47:14 crc kubenswrapper[4836]: I0122 09:47:14.915188 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.113184 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.248677 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.708892 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.760288 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.765759 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.773512 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.961948 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 22 09:47:15 crc kubenswrapper[4836]: I0122 09:47:15.977403 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.104836 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.154897 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.374689 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.477865 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.582883 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.735688 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.785144 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.933775 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 22 09:47:16 crc kubenswrapper[4836]: I0122 09:47:16.982484 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.018406 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.156487 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.158910 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.295207 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.422520 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.534857 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.626021 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.655458 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.756479 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 22 09:47:17 crc kubenswrapper[4836]: I0122 09:47:17.766872 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.044352 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.051787 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.053457 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.074307 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.254075 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.257636 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.349379 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.463436 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.498440 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.500728 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.746873 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.769534 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.802272 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 22 09:47:18 crc kubenswrapper[4836]: I0122 09:47:18.821544 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 22 09:47:19 crc kubenswrapper[4836]: I0122 09:47:19.027602 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 22 09:47:19 crc kubenswrapper[4836]: I0122 09:47:19.191462 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 22 09:47:19 crc kubenswrapper[4836]: I0122 09:47:19.396670 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 22 09:47:19 crc kubenswrapper[4836]: I0122 09:47:19.695992 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 22 09:47:19 crc kubenswrapper[4836]: I0122 09:47:19.797945 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.141991 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.184806 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.400777 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.466155 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.498535 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.608853 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.742843 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.781011 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 22 09:47:20 crc kubenswrapper[4836]: I0122 09:47:20.924828 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.049172 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.189801 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.325538 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.484537 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.514509 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.548434 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.601999 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.629951 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.652287 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.693766 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 22 09:47:21 crc kubenswrapper[4836]: I0122 09:47:21.885211 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.033779 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.094113 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.254719 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.270295 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.278493 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.290336 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.294159 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.684305 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 09:47:22 crc kubenswrapper[4836]: I0122 09:47:22.794464 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.163446 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.206378 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.453440 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.744248 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.857382 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 22 09:47:23 crc kubenswrapper[4836]: I0122 09:47:23.915536 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.412314 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.517586 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.661356 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.817624 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.845525 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.866985 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 22 09:47:24 crc kubenswrapper[4836]: I0122 09:47:24.912667 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.193406 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.204325 4836 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.217347 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.413212 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.592001 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.741825 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.789909 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.902586 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 22 09:47:25 crc kubenswrapper[4836]: I0122 09:47:25.987099 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.013484 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.014472 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.134555 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.154777 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4bdcx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.154830 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.161762 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4bdcx container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.162251 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.177212 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.432864 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.489832 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.863672 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 22 09:47:26 crc kubenswrapper[4836]: I0122 09:47:26.901293 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.012605 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.182201 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.292355 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.311595 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.442365 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.589354 4836 generic.go:334] "Generic (PLEG): container finished" podID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerID="54eb9b908456f56ece981ae792ccb16c3e8150b01e7d353f2f834607d8d287e6" exitCode=0 Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.589423 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerDied","Data":"54eb9b908456f56ece981ae792ccb16c3e8150b01e7d353f2f834607d8d287e6"} Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.590405 4836 scope.go:117] "RemoveContainer" containerID="54eb9b908456f56ece981ae792ccb16c3e8150b01e7d353f2f834607d8d287e6" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.610258 4836 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.610894 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cpx6h" podStartSLOduration=59.361134322 podStartE2EDuration="2m58.610878841s" podCreationTimestamp="2026-01-22 09:44:29 +0000 UTC" firstStartedPulling="2026-01-22 09:44:38.27417737 +0000 UTC m=+161.633945887" lastFinishedPulling="2026-01-22 09:46:37.523921889 +0000 UTC m=+280.883690406" observedRunningTime="2026-01-22 09:46:38.294396343 +0000 UTC m=+281.654164860" watchObservedRunningTime="2026-01-22 09:47:27.610878841 +0000 UTC m=+330.970647358" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.612175 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=82.612168314 podStartE2EDuration="1m22.612168314s" podCreationTimestamp="2026-01-22 09:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:36.750355012 +0000 UTC m=+280.110123549" watchObservedRunningTime="2026-01-22 09:47:27.612168314 +0000 UTC m=+330.971936831" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.616238 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.616354 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.648228 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.675456 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=51.675423405 podStartE2EDuration="51.675423405s" podCreationTimestamp="2026-01-22 09:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:47:27.656597827 +0000 UTC m=+331.016366394" watchObservedRunningTime="2026-01-22 09:47:27.675423405 +0000 UTC m=+331.035191922" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.733644 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.769611 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 22 09:47:27 crc kubenswrapper[4836]: I0122 09:47:27.852973 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.019365 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.043509 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.141551 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.142357 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.165699 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.222724 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.251942 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.273898 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.513199 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.523940 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.560177 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.761937 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 22 09:47:28 crc kubenswrapper[4836]: I0122 09:47:28.827313 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.022422 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.022464 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.028576 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.150323 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.248107 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.280901 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.489070 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.576872 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.602394 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.603232 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4bdcx_ee9368ed-7945-471a-9869-aa2fdb8ad2c7/marketplace-operator/1.log" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.603840 4836 generic.go:334] "Generic (PLEG): container finished" podID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" exitCode=1 Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.603928 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerDied","Data":"8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2"} Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.603989 4836 scope.go:117] "RemoveContainer" containerID="54eb9b908456f56ece981ae792ccb16c3e8150b01e7d353f2f834607d8d287e6" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.604810 4836 scope.go:117] "RemoveContainer" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" Jan 22 09:47:29 crc kubenswrapper[4836]: E0122 09:47:29.605164 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4bdcx_openshift-marketplace(ee9368ed-7945-471a-9869-aa2fdb8ad2c7)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.608637 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.737794 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 22 09:47:29 crc kubenswrapper[4836]: I0122 09:47:29.758312 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.023730 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.434493 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.613138 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4bdcx_ee9368ed-7945-471a-9869-aa2fdb8ad2c7/marketplace-operator/1.log" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.627750 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.660177 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.724707 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 22 09:47:30 crc kubenswrapper[4836]: I0122 09:47:30.855694 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.009719 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.079602 4836 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.275303 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.323254 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.351097 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.429362 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.525829 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.529843 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.616467 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.700768 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.782096 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.872336 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.920602 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 22 09:47:31 crc kubenswrapper[4836]: I0122 09:47:31.967463 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.062634 4836 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.063124 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2" gracePeriod=5 Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.087981 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.148104 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.191752 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.192881 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.202849 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.314701 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.464721 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.513258 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.713149 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 22 09:47:32 crc kubenswrapper[4836]: I0122 09:47:32.790385 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.243991 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.337985 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.620331 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.628458 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.664919 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 22 09:47:33 crc kubenswrapper[4836]: I0122 09:47:33.693708 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 22 09:47:34 crc kubenswrapper[4836]: I0122 09:47:34.014025 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 09:47:34 crc kubenswrapper[4836]: I0122 09:47:34.019244 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 22 09:47:34 crc kubenswrapper[4836]: I0122 09:47:34.349826 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 22 09:47:34 crc kubenswrapper[4836]: I0122 09:47:34.571365 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 22 09:47:35 crc kubenswrapper[4836]: I0122 09:47:35.043969 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 22 09:47:35 crc kubenswrapper[4836]: I0122 09:47:35.295512 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 22 09:47:35 crc kubenswrapper[4836]: I0122 09:47:35.459785 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 22 09:47:35 crc kubenswrapper[4836]: I0122 09:47:35.745303 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.153842 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.154576 4836 scope.go:117] "RemoveContainer" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" Jan 22 09:47:36 crc kubenswrapper[4836]: E0122 09:47:36.154863 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4bdcx_openshift-marketplace(ee9368ed-7945-471a-9869-aa2fdb8ad2c7)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.155200 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.402278 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.538434 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 22 09:47:36 crc kubenswrapper[4836]: I0122 09:47:36.676476 4836 scope.go:117] "RemoveContainer" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" Jan 22 09:47:36 crc kubenswrapper[4836]: E0122 09:47:36.676659 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4bdcx_openshift-marketplace(ee9368ed-7945-471a-9869-aa2fdb8ad2c7)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.310626 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.332887 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.549701 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.663534 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.663610 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.671168 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.684871 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.684935 4836 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2" exitCode=137 Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.684986 4836 scope.go:117] "RemoveContainer" containerID="4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.685005 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.708929 4836 scope.go:117] "RemoveContainer" containerID="4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2" Jan 22 09:47:37 crc kubenswrapper[4836]: E0122 09:47:37.709653 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2\": container with ID starting with 4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2 not found: ID does not exist" containerID="4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.709770 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2"} err="failed to get container status \"4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2\": rpc error: code = NotFound desc = could not find container \"4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2\": container with ID starting with 4fb6e67a56afd766e7e63b1c116cc740acc2726bf001ecfaa630eccc3fc1c9d2 not found: ID does not exist" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795100 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795186 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795304 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795384 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795433 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795463 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795519 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795557 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795658 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795921 4836 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795951 4836 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795970 4836 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.795994 4836 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.808091 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:47:37 crc kubenswrapper[4836]: I0122 09:47:37.897801 4836 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.221529 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.279214 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.370404 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.370724 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" podUID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" containerName="controller-manager" containerID="cri-o://a838fbcfa417f78c2b3a1fd03ad4732494737cdb79000727629fe38bf25c1d1c" gracePeriod=30 Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.464114 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.464363 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerName="route-controller-manager" containerID="cri-o://bf65d926788eeb3b71e09c48eed24099087fef5838a02fb17c54f914a428d11f" gracePeriod=30 Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.680368 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.696094 4836 generic.go:334] "Generic (PLEG): container finished" podID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" containerID="a838fbcfa417f78c2b3a1fd03ad4732494737cdb79000727629fe38bf25c1d1c" exitCode=0 Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.696211 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" event={"ID":"17b4ea51-6711-4c06-a17e-5a26b0590fa6","Type":"ContainerDied","Data":"a838fbcfa417f78c2b3a1fd03ad4732494737cdb79000727629fe38bf25c1d1c"} Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.700501 4836 generic.go:334] "Generic (PLEG): container finished" podID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerID="bf65d926788eeb3b71e09c48eed24099087fef5838a02fb17c54f914a428d11f" exitCode=0 Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.700565 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" event={"ID":"794eb7c1-6182-4792-9798-5553ebf10ad4","Type":"ContainerDied","Data":"bf65d926788eeb3b71e09c48eed24099087fef5838a02fb17c54f914a428d11f"} Jan 22 09:47:38 crc kubenswrapper[4836]: I0122 09:47:38.918777 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.009646 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.010864 4836 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.029253 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.029314 4836 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0ef8d8b1-9977-416d-8910-36cf1cccacd7" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.034625 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.034660 4836 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0ef8d8b1-9977-416d-8910-36cf1cccacd7" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.045020 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.213305 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca\") pod \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.213431 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert\") pod \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.213474 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles\") pod \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.213537 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config\") pod \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.213580 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d7vn\" (UniqueName: \"kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn\") pod \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\" (UID: \"17b4ea51-6711-4c06-a17e-5a26b0590fa6\") " Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.214655 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca" (OuterVolumeSpecName: "client-ca") pod "17b4ea51-6711-4c06-a17e-5a26b0590fa6" (UID: "17b4ea51-6711-4c06-a17e-5a26b0590fa6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.214729 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "17b4ea51-6711-4c06-a17e-5a26b0590fa6" (UID: "17b4ea51-6711-4c06-a17e-5a26b0590fa6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.214939 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config" (OuterVolumeSpecName: "config") pod "17b4ea51-6711-4c06-a17e-5a26b0590fa6" (UID: "17b4ea51-6711-4c06-a17e-5a26b0590fa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.220717 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn" (OuterVolumeSpecName: "kube-api-access-7d7vn") pod "17b4ea51-6711-4c06-a17e-5a26b0590fa6" (UID: "17b4ea51-6711-4c06-a17e-5a26b0590fa6"). InnerVolumeSpecName "kube-api-access-7d7vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.220736 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "17b4ea51-6711-4c06-a17e-5a26b0590fa6" (UID: "17b4ea51-6711-4c06-a17e-5a26b0590fa6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.315445 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.315501 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d7vn\" (UniqueName: \"kubernetes.io/projected/17b4ea51-6711-4c06-a17e-5a26b0590fa6-kube-api-access-7d7vn\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.315524 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.315546 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17b4ea51-6711-4c06-a17e-5a26b0590fa6-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.315564 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/17b4ea51-6711-4c06-a17e-5a26b0590fa6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.360729 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.396953 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.508032 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.704095 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.709033 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" event={"ID":"17b4ea51-6711-4c06-a17e-5a26b0590fa6","Type":"ContainerDied","Data":"3d054b69866190511ffb2fdbb38bd52c29e09552413568de26daae6a64ad4029"} Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.709102 4836 scope.go:117] "RemoveContainer" containerID="a838fbcfa417f78c2b3a1fd03ad4732494737cdb79000727629fe38bf25c1d1c" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.709229 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d48677986-ttmcf" Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.744752 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.748481 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-d48677986-ttmcf"] Jan 22 09:47:39 crc kubenswrapper[4836]: I0122 09:47:39.916352 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.025254 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config\") pod \"794eb7c1-6182-4792-9798-5553ebf10ad4\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.025371 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvsz5\" (UniqueName: \"kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5\") pod \"794eb7c1-6182-4792-9798-5553ebf10ad4\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.025397 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert\") pod \"794eb7c1-6182-4792-9798-5553ebf10ad4\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.025458 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca\") pod \"794eb7c1-6182-4792-9798-5553ebf10ad4\" (UID: \"794eb7c1-6182-4792-9798-5553ebf10ad4\") " Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.026381 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca" (OuterVolumeSpecName: "client-ca") pod "794eb7c1-6182-4792-9798-5553ebf10ad4" (UID: "794eb7c1-6182-4792-9798-5553ebf10ad4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.026471 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config" (OuterVolumeSpecName: "config") pod "794eb7c1-6182-4792-9798-5553ebf10ad4" (UID: "794eb7c1-6182-4792-9798-5553ebf10ad4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.030924 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "794eb7c1-6182-4792-9798-5553ebf10ad4" (UID: "794eb7c1-6182-4792-9798-5553ebf10ad4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.031169 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5" (OuterVolumeSpecName: "kube-api-access-xvsz5") pod "794eb7c1-6182-4792-9798-5553ebf10ad4" (UID: "794eb7c1-6182-4792-9798-5553ebf10ad4"). InnerVolumeSpecName "kube-api-access-xvsz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.068489 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:47:40 crc kubenswrapper[4836]: E0122 09:47:40.068915 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" containerName="installer" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.068946 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" containerName="installer" Jan 22 09:47:40 crc kubenswrapper[4836]: E0122 09:47:40.068965 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerName="route-controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.068977 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerName="route-controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: E0122 09:47:40.068996 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069010 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 09:47:40 crc kubenswrapper[4836]: E0122 09:47:40.069036 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" containerName="controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069072 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" containerName="controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069263 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" containerName="route-controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069309 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" containerName="controller-manager" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069334 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.069352 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b6b777-0f50-4dd2-af16-1f27e8969a7f" containerName="installer" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.070019 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.072864 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.073447 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.073776 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.073801 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.075009 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.076170 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.076394 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.079495 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.085034 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.092153 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.094533 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.126516 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.126572 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvsz5\" (UniqueName: \"kubernetes.io/projected/794eb7c1-6182-4792-9798-5553ebf10ad4-kube-api-access-xvsz5\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.126586 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/794eb7c1-6182-4792-9798-5553ebf10ad4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.126600 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/794eb7c1-6182-4792-9798-5553ebf10ad4-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.170359 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.228572 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.228679 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.228734 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.228829 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.228864 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.229015 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.229208 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76pjj\" (UniqueName: \"kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.229277 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8ld5\" (UniqueName: \"kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.229451 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331107 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331234 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331277 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331335 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8ld5\" (UniqueName: \"kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331377 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76pjj\" (UniqueName: \"kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331436 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331505 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331546 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.331591 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.332892 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.333581 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.334189 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.334506 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.335586 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.339083 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.339205 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.361067 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76pjj\" (UniqueName: \"kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj\") pod \"route-controller-manager-6997447f6-9jml9\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.362128 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8ld5\" (UniqueName: \"kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5\") pod \"controller-manager-685dd7c657-2md8m\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.387903 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.395655 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.502772 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.646705 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.714548 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.718158 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.718154 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc" event={"ID":"794eb7c1-6182-4792-9798-5553ebf10ad4","Type":"ContainerDied","Data":"faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a"} Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.718239 4836 scope.go:117] "RemoveContainer" containerID="bf65d926788eeb3b71e09c48eed24099087fef5838a02fb17c54f914a428d11f" Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.725860 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" event={"ID":"853016e1-85bd-4bf3-a963-8982e9decd95","Type":"ContainerStarted","Data":"793bf5abcf5fcee1e68a537377ffc8992dbb4b473a600ce881f663659a833f57"} Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.789976 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.793055 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d9fb8b77c-bdmvc"] Jan 22 09:47:40 crc kubenswrapper[4836]: I0122 09:47:40.847139 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 22 09:47:40 crc kubenswrapper[4836]: E0122 09:47:40.888008 4836 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod794eb7c1_6182_4792_9798_5553ebf10ad4.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod794eb7c1_6182_4792_9798_5553ebf10ad4.slice/crio-faa4d70d28a804ee5aaf03d5473dc5604ac56b26b26c6673298272f1cfe9e00a\": RecentStats: unable to find data in memory cache]" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.008376 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17b4ea51-6711-4c06-a17e-5a26b0590fa6" path="/var/lib/kubelet/pods/17b4ea51-6711-4c06-a17e-5a26b0590fa6/volumes" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.009668 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="794eb7c1-6182-4792-9798-5553ebf10ad4" path="/var/lib/kubelet/pods/794eb7c1-6182-4792-9798-5553ebf10ad4/volumes" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.266940 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.435283 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.618443 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.741172 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" event={"ID":"e6bfe8cd-bae5-48c7-80ef-165c0181720f","Type":"ContainerStarted","Data":"b32098012666f418005ca2b6950fac2c737e8e9509cb58abd208d2083e270afd"} Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.741227 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" event={"ID":"e6bfe8cd-bae5-48c7-80ef-165c0181720f","Type":"ContainerStarted","Data":"0b7622646c796f1342d7c298584cc35a47b01028d71876abb20627a3c6212449"} Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.741564 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.744189 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" event={"ID":"853016e1-85bd-4bf3-a963-8982e9decd95","Type":"ContainerStarted","Data":"fdf60fc6ed6a3b28b8174e6d1e0b98757881773316299c292038db362b8fd0a3"} Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.744432 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.748613 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.749712 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.765586 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" podStartSLOduration=3.7655648360000002 podStartE2EDuration="3.765564836s" podCreationTimestamp="2026-01-22 09:47:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:47:41.761822609 +0000 UTC m=+345.121591186" watchObservedRunningTime="2026-01-22 09:47:41.765564836 +0000 UTC m=+345.125333393" Jan 22 09:47:41 crc kubenswrapper[4836]: I0122 09:47:41.795453 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" podStartSLOduration=3.79542456 podStartE2EDuration="3.79542456s" podCreationTimestamp="2026-01-22 09:47:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:47:41.784256821 +0000 UTC m=+345.144025368" watchObservedRunningTime="2026-01-22 09:47:41.79542456 +0000 UTC m=+345.155193107" Jan 22 09:47:48 crc kubenswrapper[4836]: I0122 09:47:48.996825 4836 scope.go:117] "RemoveContainer" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" Jan 22 09:47:51 crc kubenswrapper[4836]: I0122 09:47:51.814197 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4bdcx_ee9368ed-7945-471a-9869-aa2fdb8ad2c7/marketplace-operator/1.log" Jan 22 09:47:51 crc kubenswrapper[4836]: I0122 09:47:51.814670 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerStarted","Data":"a34725fae5b31b25a83b1721699de5384d4639899d3b350b04caeec07e3f1329"} Jan 22 09:47:51 crc kubenswrapper[4836]: I0122 09:47:51.815002 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:47:51 crc kubenswrapper[4836]: I0122 09:47:51.818025 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:47:53 crc kubenswrapper[4836]: I0122 09:47:53.897913 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:47:53 crc kubenswrapper[4836]: I0122 09:47:53.898413 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.348807 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.350497 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" podUID="853016e1-85bd-4bf3-a963-8982e9decd95" containerName="controller-manager" containerID="cri-o://fdf60fc6ed6a3b28b8174e6d1e0b98757881773316299c292038db362b8fd0a3" gracePeriod=30 Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.365064 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.365595 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" podUID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" containerName="route-controller-manager" containerID="cri-o://b32098012666f418005ca2b6950fac2c737e8e9509cb58abd208d2083e270afd" gracePeriod=30 Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.855101 4836 generic.go:334] "Generic (PLEG): container finished" podID="853016e1-85bd-4bf3-a963-8982e9decd95" containerID="fdf60fc6ed6a3b28b8174e6d1e0b98757881773316299c292038db362b8fd0a3" exitCode=0 Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.855176 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" event={"ID":"853016e1-85bd-4bf3-a963-8982e9decd95","Type":"ContainerDied","Data":"fdf60fc6ed6a3b28b8174e6d1e0b98757881773316299c292038db362b8fd0a3"} Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.857525 4836 generic.go:334] "Generic (PLEG): container finished" podID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" containerID="b32098012666f418005ca2b6950fac2c737e8e9509cb58abd208d2083e270afd" exitCode=0 Jan 22 09:47:58 crc kubenswrapper[4836]: I0122 09:47:58.857579 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" event={"ID":"e6bfe8cd-bae5-48c7-80ef-165c0181720f","Type":"ContainerDied","Data":"b32098012666f418005ca2b6950fac2c737e8e9509cb58abd208d2083e270afd"} Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.492323 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.530191 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:47:59 crc kubenswrapper[4836]: E0122 09:47:59.531142 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" containerName="route-controller-manager" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.531287 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" containerName="route-controller-manager" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.531602 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" containerName="route-controller-manager" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.532798 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.542461 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.559720 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.605833 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert\") pod \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.605924 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config\") pod \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606004 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76pjj\" (UniqueName: \"kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj\") pod \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606068 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca\") pod \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\" (UID: \"e6bfe8cd-bae5-48c7-80ef-165c0181720f\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606711 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca" (OuterVolumeSpecName: "client-ca") pod "e6bfe8cd-bae5-48c7-80ef-165c0181720f" (UID: "e6bfe8cd-bae5-48c7-80ef-165c0181720f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606747 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config" (OuterVolumeSpecName: "config") pod "e6bfe8cd-bae5-48c7-80ef-165c0181720f" (UID: "e6bfe8cd-bae5-48c7-80ef-165c0181720f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606881 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606947 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4bpg\" (UniqueName: \"kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.606996 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.607014 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.607084 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.607102 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bfe8cd-bae5-48c7-80ef-165c0181720f-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.611516 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e6bfe8cd-bae5-48c7-80ef-165c0181720f" (UID: "e6bfe8cd-bae5-48c7-80ef-165c0181720f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.612725 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj" (OuterVolumeSpecName: "kube-api-access-76pjj") pod "e6bfe8cd-bae5-48c7-80ef-165c0181720f" (UID: "e6bfe8cd-bae5-48c7-80ef-165c0181720f"). InnerVolumeSpecName "kube-api-access-76pjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.707964 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles\") pod \"853016e1-85bd-4bf3-a963-8982e9decd95\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.708311 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert\") pod \"853016e1-85bd-4bf3-a963-8982e9decd95\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.708445 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca\") pod \"853016e1-85bd-4bf3-a963-8982e9decd95\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.708945 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "853016e1-85bd-4bf3-a963-8982e9decd95" (UID: "853016e1-85bd-4bf3-a963-8982e9decd95"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.709014 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca" (OuterVolumeSpecName: "client-ca") pod "853016e1-85bd-4bf3-a963-8982e9decd95" (UID: "853016e1-85bd-4bf3-a963-8982e9decd95"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.709204 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config\") pod \"853016e1-85bd-4bf3-a963-8982e9decd95\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.709686 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8ld5\" (UniqueName: \"kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5\") pod \"853016e1-85bd-4bf3-a963-8982e9decd95\" (UID: \"853016e1-85bd-4bf3-a963-8982e9decd95\") " Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.709922 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config" (OuterVolumeSpecName: "config") pod "853016e1-85bd-4bf3-a963-8982e9decd95" (UID: "853016e1-85bd-4bf3-a963-8982e9decd95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710115 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710242 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710403 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710541 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4bpg\" (UniqueName: \"kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710656 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710745 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76pjj\" (UniqueName: \"kubernetes.io/projected/e6bfe8cd-bae5-48c7-80ef-165c0181720f-kube-api-access-76pjj\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710829 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710903 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/853016e1-85bd-4bf3-a963-8982e9decd95-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.710980 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bfe8cd-bae5-48c7-80ef-165c0181720f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.711618 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "853016e1-85bd-4bf3-a963-8982e9decd95" (UID: "853016e1-85bd-4bf3-a963-8982e9decd95"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.711791 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.711949 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5" (OuterVolumeSpecName: "kube-api-access-v8ld5") pod "853016e1-85bd-4bf3-a963-8982e9decd95" (UID: "853016e1-85bd-4bf3-a963-8982e9decd95"). InnerVolumeSpecName "kube-api-access-v8ld5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.712099 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.714591 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.731744 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4bpg\" (UniqueName: \"kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg\") pod \"route-controller-manager-5bdf786554-4sv4x\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.812144 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8ld5\" (UniqueName: \"kubernetes.io/projected/853016e1-85bd-4bf3-a963-8982e9decd95-kube-api-access-v8ld5\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.812207 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/853016e1-85bd-4bf3-a963-8982e9decd95-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.865964 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" event={"ID":"853016e1-85bd-4bf3-a963-8982e9decd95","Type":"ContainerDied","Data":"793bf5abcf5fcee1e68a537377ffc8992dbb4b473a600ce881f663659a833f57"} Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.866009 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-2md8m" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.866343 4836 scope.go:117] "RemoveContainer" containerID="fdf60fc6ed6a3b28b8174e6d1e0b98757881773316299c292038db362b8fd0a3" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.869135 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.869162 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9" event={"ID":"e6bfe8cd-bae5-48c7-80ef-165c0181720f","Type":"ContainerDied","Data":"0b7622646c796f1342d7c298584cc35a47b01028d71876abb20627a3c6212449"} Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.870388 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.908233 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.913711 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-9jml9"] Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.925432 4836 scope.go:117] "RemoveContainer" containerID="b32098012666f418005ca2b6950fac2c737e8e9509cb58abd208d2083e270afd" Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.950199 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:47:59 crc kubenswrapper[4836]: I0122 09:47:59.955427 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-2md8m"] Jan 22 09:48:00 crc kubenswrapper[4836]: I0122 09:48:00.316000 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:48:00 crc kubenswrapper[4836]: W0122 09:48:00.320499 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod323eb4bf_e470_4ac2_801e_9150a4afdb5e.slice/crio-d6fa4580fa06b56760da68adf1a16b7fc08aa7ed0782e6104fe508f9d3ee3c92 WatchSource:0}: Error finding container d6fa4580fa06b56760da68adf1a16b7fc08aa7ed0782e6104fe508f9d3ee3c92: Status 404 returned error can't find the container with id d6fa4580fa06b56760da68adf1a16b7fc08aa7ed0782e6104fe508f9d3ee3c92 Jan 22 09:48:00 crc kubenswrapper[4836]: I0122 09:48:00.880004 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" event={"ID":"323eb4bf-e470-4ac2-801e-9150a4afdb5e","Type":"ContainerStarted","Data":"91020b0963240054e78b6269ceed248299570bc059636ecb5c0aad0b63014933"} Jan 22 09:48:00 crc kubenswrapper[4836]: I0122 09:48:00.880375 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" event={"ID":"323eb4bf-e470-4ac2-801e-9150a4afdb5e","Type":"ContainerStarted","Data":"d6fa4580fa06b56760da68adf1a16b7fc08aa7ed0782e6104fe508f9d3ee3c92"} Jan 22 09:48:00 crc kubenswrapper[4836]: I0122 09:48:00.880482 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:48:00 crc kubenswrapper[4836]: I0122 09:48:00.905365 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" podStartSLOduration=2.905346559 podStartE2EDuration="2.905346559s" podCreationTimestamp="2026-01-22 09:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:00.902009473 +0000 UTC m=+364.261778050" watchObservedRunningTime="2026-01-22 09:48:00.905346559 +0000 UTC m=+364.265115076" Jan 22 09:48:01 crc kubenswrapper[4836]: I0122 09:48:01.006232 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="853016e1-85bd-4bf3-a963-8982e9decd95" path="/var/lib/kubelet/pods/853016e1-85bd-4bf3-a963-8982e9decd95/volumes" Jan 22 09:48:01 crc kubenswrapper[4836]: I0122 09:48:01.007014 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6bfe8cd-bae5-48c7-80ef-165c0181720f" path="/var/lib/kubelet/pods/e6bfe8cd-bae5-48c7-80ef-165c0181720f/volumes" Jan 22 09:48:01 crc kubenswrapper[4836]: I0122 09:48:01.180484 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.082310 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:48:02 crc kubenswrapper[4836]: E0122 09:48:02.082643 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853016e1-85bd-4bf3-a963-8982e9decd95" containerName="controller-manager" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.082664 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="853016e1-85bd-4bf3-a963-8982e9decd95" containerName="controller-manager" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.082841 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="853016e1-85bd-4bf3-a963-8982e9decd95" containerName="controller-manager" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.083453 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.095664 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.096274 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.096797 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.098995 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.100552 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.110005 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.123012 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.131716 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.142722 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmghr\" (UniqueName: \"kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.142784 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.142824 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.142914 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.142939 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.244182 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmghr\" (UniqueName: \"kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.244840 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.245004 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.245185 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.245327 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.246094 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.246171 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.246986 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.256232 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.261280 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmghr\" (UniqueName: \"kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr\") pod \"controller-manager-7c45df54bf-gqz5t\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.425752 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.821175 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:48:02 crc kubenswrapper[4836]: W0122 09:48:02.833375 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a59c7e5_71fe_4591_ba35_00e94ee8ba6d.slice/crio-ba5c8219c40f16b7817bf5e8a2b73c3bc604bda9988b4fd222df6f5f1d9024f3 WatchSource:0}: Error finding container ba5c8219c40f16b7817bf5e8a2b73c3bc604bda9988b4fd222df6f5f1d9024f3: Status 404 returned error can't find the container with id ba5c8219c40f16b7817bf5e8a2b73c3bc604bda9988b4fd222df6f5f1d9024f3 Jan 22 09:48:02 crc kubenswrapper[4836]: I0122 09:48:02.890924 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" event={"ID":"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d","Type":"ContainerStarted","Data":"ba5c8219c40f16b7817bf5e8a2b73c3bc604bda9988b4fd222df6f5f1d9024f3"} Jan 22 09:48:03 crc kubenswrapper[4836]: I0122 09:48:03.899781 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" event={"ID":"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d","Type":"ContainerStarted","Data":"99b7942769646bafd8d1a4d63f589e249a9d0fe5366c6936afb10bde148141cd"} Jan 22 09:48:03 crc kubenswrapper[4836]: I0122 09:48:03.901129 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:03 crc kubenswrapper[4836]: I0122 09:48:03.909952 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:03 crc kubenswrapper[4836]: I0122 09:48:03.930288 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" podStartSLOduration=5.930269676 podStartE2EDuration="5.930269676s" podCreationTimestamp="2026-01-22 09:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:03.926702803 +0000 UTC m=+367.286471360" watchObservedRunningTime="2026-01-22 09:48:03.930269676 +0000 UTC m=+367.290038203" Jan 22 09:48:23 crc kubenswrapper[4836]: I0122 09:48:23.897584 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:48:23 crc kubenswrapper[4836]: I0122 09:48:23.898311 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:48:43 crc kubenswrapper[4836]: I0122 09:48:43.861159 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:48:43 crc kubenswrapper[4836]: I0122 09:48:43.862995 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zbvh5" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="registry-server" containerID="cri-o://491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c" gracePeriod=2 Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.066005 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.159386 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n4dwm" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="registry-server" containerID="cri-o://cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29" gracePeriod=2 Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.659652 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.714424 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.768530 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvznw\" (UniqueName: \"kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw\") pod \"1b2b98a3-9481-442f-a35a-d8adcf779887\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.768615 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities\") pod \"1b2b98a3-9481-442f-a35a-d8adcf779887\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.768655 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content\") pod \"1b2b98a3-9481-442f-a35a-d8adcf779887\" (UID: \"1b2b98a3-9481-442f-a35a-d8adcf779887\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.769738 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities" (OuterVolumeSpecName: "utilities") pod "1b2b98a3-9481-442f-a35a-d8adcf779887" (UID: "1b2b98a3-9481-442f-a35a-d8adcf779887"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.775076 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw" (OuterVolumeSpecName: "kube-api-access-mvznw") pod "1b2b98a3-9481-442f-a35a-d8adcf779887" (UID: "1b2b98a3-9481-442f-a35a-d8adcf779887"). InnerVolumeSpecName "kube-api-access-mvznw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.821880 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b2b98a3-9481-442f-a35a-d8adcf779887" (UID: "1b2b98a3-9481-442f-a35a-d8adcf779887"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870058 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content\") pod \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870292 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbnjd\" (UniqueName: \"kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd\") pod \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870359 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities\") pod \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\" (UID: \"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8\") " Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870770 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvznw\" (UniqueName: \"kubernetes.io/projected/1b2b98a3-9481-442f-a35a-d8adcf779887-kube-api-access-mvznw\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870788 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.870800 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b2b98a3-9481-442f-a35a-d8adcf779887-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.871139 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities" (OuterVolumeSpecName: "utilities") pod "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" (UID: "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.873150 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd" (OuterVolumeSpecName: "kube-api-access-xbnjd") pod "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" (UID: "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8"). InnerVolumeSpecName "kube-api-access-xbnjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.921688 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" (UID: "9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.971934 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.971983 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbnjd\" (UniqueName: \"kubernetes.io/projected/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-kube-api-access-xbnjd\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:44 crc kubenswrapper[4836]: I0122 09:48:44.971995 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.171302 4836 generic.go:334] "Generic (PLEG): container finished" podID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerID="491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c" exitCode=0 Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.171455 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerDied","Data":"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c"} Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.171514 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbvh5" event={"ID":"9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8","Type":"ContainerDied","Data":"503f1b58afd615dc10a0198db457c9a8307f810f69e8c8d41f7bab0de10be715"} Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.171548 4836 scope.go:117] "RemoveContainer" containerID="491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.171763 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbvh5" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.182785 4836 generic.go:334] "Generic (PLEG): container finished" podID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerID="cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29" exitCode=0 Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.182905 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerDied","Data":"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29"} Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.182927 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n4dwm" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.182966 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n4dwm" event={"ID":"1b2b98a3-9481-442f-a35a-d8adcf779887","Type":"ContainerDied","Data":"1eab1bac3d9313405c4a355ebc4f696142dce4b82041457bd2e5fb9ae03c1142"} Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.213487 4836 scope.go:117] "RemoveContainer" containerID="80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.222109 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.227667 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zbvh5"] Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.236333 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.241069 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n4dwm"] Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.243634 4836 scope.go:117] "RemoveContainer" containerID="e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.261367 4836 scope.go:117] "RemoveContainer" containerID="491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.262014 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c\": container with ID starting with 491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c not found: ID does not exist" containerID="491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.262107 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c"} err="failed to get container status \"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c\": rpc error: code = NotFound desc = could not find container \"491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c\": container with ID starting with 491b353be8de8666d2ebef2cc573a1b97c7d46471071efa1f68834f3d7c52a1c not found: ID does not exist" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.262154 4836 scope.go:117] "RemoveContainer" containerID="80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.262597 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6\": container with ID starting with 80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6 not found: ID does not exist" containerID="80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.262640 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6"} err="failed to get container status \"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6\": rpc error: code = NotFound desc = could not find container \"80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6\": container with ID starting with 80476968a1d411c56db9b13397eb2deed600770c0dfb266da2c8012eba71f7d6 not found: ID does not exist" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.262669 4836 scope.go:117] "RemoveContainer" containerID="e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.262960 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106\": container with ID starting with e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106 not found: ID does not exist" containerID="e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.263003 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106"} err="failed to get container status \"e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106\": rpc error: code = NotFound desc = could not find container \"e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106\": container with ID starting with e345c7e822d6fe816570466b88105e9121b9f7a1083eeaa9d753d269e28fe106 not found: ID does not exist" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.263029 4836 scope.go:117] "RemoveContainer" containerID="cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.277647 4836 scope.go:117] "RemoveContainer" containerID="66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.293006 4836 scope.go:117] "RemoveContainer" containerID="2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.312075 4836 scope.go:117] "RemoveContainer" containerID="cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.313711 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29\": container with ID starting with cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29 not found: ID does not exist" containerID="cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.313768 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29"} err="failed to get container status \"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29\": rpc error: code = NotFound desc = could not find container \"cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29\": container with ID starting with cd3dca208d5d22e990d1018b5233ea2db3ee35496143be560cf7735ce71c7c29 not found: ID does not exist" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.313809 4836 scope.go:117] "RemoveContainer" containerID="66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.314263 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6\": container with ID starting with 66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6 not found: ID does not exist" containerID="66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.314299 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6"} err="failed to get container status \"66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6\": rpc error: code = NotFound desc = could not find container \"66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6\": container with ID starting with 66ec0ba40cd5947807af8dcebf67aa8108d97657f960a8f7f55f85baee5d0de6 not found: ID does not exist" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.314328 4836 scope.go:117] "RemoveContainer" containerID="2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673" Jan 22 09:48:45 crc kubenswrapper[4836]: E0122 09:48:45.314741 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673\": container with ID starting with 2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673 not found: ID does not exist" containerID="2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673" Jan 22 09:48:45 crc kubenswrapper[4836]: I0122 09:48:45.314768 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673"} err="failed to get container status \"2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673\": rpc error: code = NotFound desc = could not find container \"2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673\": container with ID starting with 2dac0d76b8bf995c2df435dd0695cc1cff64792dea499a01d4acb378673f6673 not found: ID does not exist" Jan 22 09:48:46 crc kubenswrapper[4836]: I0122 09:48:46.265793 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:48:46 crc kubenswrapper[4836]: I0122 09:48:46.266607 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-86vhq" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="registry-server" containerID="cri-o://9e2f2bfeec75c92b16687fdafb4ea62255a7150d193bcd852a5117c2990ad173" gracePeriod=2 Jan 22 09:48:46 crc kubenswrapper[4836]: I0122 09:48:46.461453 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:48:46 crc kubenswrapper[4836]: I0122 09:48:46.461669 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2xnzx" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="registry-server" containerID="cri-o://ef23fc6e1e30992548a6a333e18e8d1c11bed5e79ee0d6524a9ed3610e2d6954" gracePeriod=2 Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.004201 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" path="/var/lib/kubelet/pods/1b2b98a3-9481-442f-a35a-d8adcf779887/volumes" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.005937 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" path="/var/lib/kubelet/pods/9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8/volumes" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.204104 4836 generic.go:334] "Generic (PLEG): container finished" podID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerID="ef23fc6e1e30992548a6a333e18e8d1c11bed5e79ee0d6524a9ed3610e2d6954" exitCode=0 Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.204166 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerDied","Data":"ef23fc6e1e30992548a6a333e18e8d1c11bed5e79ee0d6524a9ed3610e2d6954"} Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.205992 4836 generic.go:334] "Generic (PLEG): container finished" podID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerID="9e2f2bfeec75c92b16687fdafb4ea62255a7150d193bcd852a5117c2990ad173" exitCode=0 Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.206011 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerDied","Data":"9e2f2bfeec75c92b16687fdafb4ea62255a7150d193bcd852a5117c2990ad173"} Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.451705 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.510591 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607351 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content\") pod \"69209694-e7d1-4e15-bb51-2807be65ed0a\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607402 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb99t\" (UniqueName: \"kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t\") pod \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607445 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content\") pod \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607481 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nwsz\" (UniqueName: \"kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz\") pod \"69209694-e7d1-4e15-bb51-2807be65ed0a\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607524 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities\") pod \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\" (UID: \"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.607550 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities\") pod \"69209694-e7d1-4e15-bb51-2807be65ed0a\" (UID: \"69209694-e7d1-4e15-bb51-2807be65ed0a\") " Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.608397 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities" (OuterVolumeSpecName: "utilities") pod "69209694-e7d1-4e15-bb51-2807be65ed0a" (UID: "69209694-e7d1-4e15-bb51-2807be65ed0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.609152 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities" (OuterVolumeSpecName: "utilities") pod "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" (UID: "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.613966 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz" (OuterVolumeSpecName: "kube-api-access-8nwsz") pod "69209694-e7d1-4e15-bb51-2807be65ed0a" (UID: "69209694-e7d1-4e15-bb51-2807be65ed0a"). InnerVolumeSpecName "kube-api-access-8nwsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.614000 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t" (OuterVolumeSpecName: "kube-api-access-hb99t") pod "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" (UID: "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0"). InnerVolumeSpecName "kube-api-access-hb99t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.631088 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" (UID: "4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.709187 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb99t\" (UniqueName: \"kubernetes.io/projected/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-kube-api-access-hb99t\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.709236 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.709258 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nwsz\" (UniqueName: \"kubernetes.io/projected/69209694-e7d1-4e15-bb51-2807be65ed0a-kube-api-access-8nwsz\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.709275 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.709294 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.737196 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69209694-e7d1-4e15-bb51-2807be65ed0a" (UID: "69209694-e7d1-4e15-bb51-2807be65ed0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:48:47 crc kubenswrapper[4836]: I0122 09:48:47.810791 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69209694-e7d1-4e15-bb51-2807be65ed0a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.213246 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2xnzx" event={"ID":"69209694-e7d1-4e15-bb51-2807be65ed0a","Type":"ContainerDied","Data":"473a1d29c9c6fe49dac4e0475a7e2f09aa00df5d0a6d7fa0e7ae86566dcf688d"} Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.213304 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2xnzx" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.213309 4836 scope.go:117] "RemoveContainer" containerID="ef23fc6e1e30992548a6a333e18e8d1c11bed5e79ee0d6524a9ed3610e2d6954" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.216908 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86vhq" event={"ID":"4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0","Type":"ContainerDied","Data":"f2d91f91b5ab938811a79ec35bb1290582519eab5ea5fbd5f40ca2efe589a2b8"} Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.217010 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86vhq" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.237063 4836 scope.go:117] "RemoveContainer" containerID="165a17f00de4a0ed6b515b23e96577d2155575321e967ebaec84c5617f85ea46" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.240140 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.243206 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2xnzx"] Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.250700 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.253976 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-86vhq"] Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.272317 4836 scope.go:117] "RemoveContainer" containerID="8d5d71d53566bb003146644f8b6e4078ab590c33329b29b061308206ec9b567a" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.283421 4836 scope.go:117] "RemoveContainer" containerID="9e2f2bfeec75c92b16687fdafb4ea62255a7150d193bcd852a5117c2990ad173" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.294650 4836 scope.go:117] "RemoveContainer" containerID="acb68c99287e899201287e5437914995784480f2adc37e70d8d4caa5abddeb1a" Jan 22 09:48:48 crc kubenswrapper[4836]: I0122 09:48:48.305610 4836 scope.go:117] "RemoveContainer" containerID="fb304af5c031253133d5b8b38bd1753f8c670fbdb2b5f0f924f3a99a12633edf" Jan 22 09:48:49 crc kubenswrapper[4836]: I0122 09:48:49.004280 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" path="/var/lib/kubelet/pods/4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0/volumes" Jan 22 09:48:49 crc kubenswrapper[4836]: I0122 09:48:49.005164 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" path="/var/lib/kubelet/pods/69209694-e7d1-4e15-bb51-2807be65ed0a/volumes" Jan 22 09:48:53 crc kubenswrapper[4836]: I0122 09:48:53.897544 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:48:53 crc kubenswrapper[4836]: I0122 09:48:53.898768 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:48:53 crc kubenswrapper[4836]: I0122 09:48:53.898825 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:48:53 crc kubenswrapper[4836]: I0122 09:48:53.899488 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 09:48:53 crc kubenswrapper[4836]: I0122 09:48:53.899542 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86" gracePeriod=600 Jan 22 09:48:55 crc kubenswrapper[4836]: I0122 09:48:55.261472 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86" exitCode=0 Jan 22 09:48:55 crc kubenswrapper[4836]: I0122 09:48:55.261618 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86"} Jan 22 09:48:55 crc kubenswrapper[4836]: I0122 09:48:55.261848 4836 scope.go:117] "RemoveContainer" containerID="24aaf2d587f9c9ed3f5962c23851c20a23a5c96bf1f3fbde5af53f404432118f" Jan 22 09:48:56 crc kubenswrapper[4836]: I0122 09:48:56.270339 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401"} Jan 22 09:48:58 crc kubenswrapper[4836]: I0122 09:48:58.359435 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:48:58 crc kubenswrapper[4836]: I0122 09:48:58.359993 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" podUID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" containerName="controller-manager" containerID="cri-o://99b7942769646bafd8d1a4d63f589e249a9d0fe5366c6936afb10bde148141cd" gracePeriod=30 Jan 22 09:48:58 crc kubenswrapper[4836]: I0122 09:48:58.386389 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:48:58 crc kubenswrapper[4836]: I0122 09:48:58.386810 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" podUID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" containerName="route-controller-manager" containerID="cri-o://91020b0963240054e78b6269ceed248299570bc059636ecb5c0aad0b63014933" gracePeriod=30 Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.289918 4836 generic.go:334] "Generic (PLEG): container finished" podID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" containerID="99b7942769646bafd8d1a4d63f589e249a9d0fe5366c6936afb10bde148141cd" exitCode=0 Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.289990 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" event={"ID":"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d","Type":"ContainerDied","Data":"99b7942769646bafd8d1a4d63f589e249a9d0fe5366c6936afb10bde148141cd"} Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.291533 4836 generic.go:334] "Generic (PLEG): container finished" podID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" containerID="91020b0963240054e78b6269ceed248299570bc059636ecb5c0aad0b63014933" exitCode=0 Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.291581 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" event={"ID":"323eb4bf-e470-4ac2-801e-9150a4afdb5e","Type":"ContainerDied","Data":"91020b0963240054e78b6269ceed248299570bc059636ecb5c0aad0b63014933"} Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.505330 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.545837 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln"] Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546120 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546179 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546194 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546200 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546216 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546224 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546233 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546239 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546251 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546257 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546263 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546269 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546279 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546286 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546293 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546301 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546309 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546315 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546324 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546329 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546336 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" containerName="route-controller-manager" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546342 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" containerName="route-controller-manager" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546350 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546356 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="extract-utilities" Jan 22 09:48:59 crc kubenswrapper[4836]: E0122 09:48:59.546365 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546371 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="extract-content" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546460 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" containerName="route-controller-manager" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546473 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="4969f850-dcc4-4ebf-b3cf-9a59e7f8f0d0" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546480 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbcf6f5-ed55-4302-bdf6-ccf3c21a43c8" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546490 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="69209694-e7d1-4e15-bb51-2807be65ed0a" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546497 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2b98a3-9481-442f-a35a-d8adcf779887" containerName="registry-server" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.546989 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.550476 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln"] Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.558128 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665522 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config\") pod \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665581 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config\") pod \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665630 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca\") pod \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665693 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles\") pod \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665723 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4bpg\" (UniqueName: \"kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg\") pod \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665787 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert\") pod \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665831 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert\") pod \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.665856 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca\") pod \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\" (UID: \"323eb4bf-e470-4ac2-801e-9150a4afdb5e\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666692 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca" (OuterVolumeSpecName: "client-ca") pod "323eb4bf-e470-4ac2-801e-9150a4afdb5e" (UID: "323eb4bf-e470-4ac2-801e-9150a4afdb5e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666737 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config" (OuterVolumeSpecName: "config") pod "323eb4bf-e470-4ac2-801e-9150a4afdb5e" (UID: "323eb4bf-e470-4ac2-801e-9150a4afdb5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666707 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca" (OuterVolumeSpecName: "client-ca") pod "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" (UID: "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666330 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmghr\" (UniqueName: \"kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr\") pod \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\" (UID: \"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d\") " Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666809 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" (UID: "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666834 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config" (OuterVolumeSpecName: "config") pod "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" (UID: "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.666950 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-client-ca\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667025 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfd2t\" (UniqueName: \"kubernetes.io/projected/dd7a889f-a301-4b2e-a471-8b481a724509-kube-api-access-zfd2t\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667209 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-config\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667267 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7a889f-a301-4b2e-a471-8b481a724509-serving-cert\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667392 4836 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667436 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667454 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/323eb4bf-e470-4ac2-801e-9150a4afdb5e-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667470 4836 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.667485 4836 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-client-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.670726 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr" (OuterVolumeSpecName: "kube-api-access-vmghr") pod "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" (UID: "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d"). InnerVolumeSpecName "kube-api-access-vmghr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.671058 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" (UID: "0a59c7e5-71fe-4591-ba35-00e94ee8ba6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.671465 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "323eb4bf-e470-4ac2-801e-9150a4afdb5e" (UID: "323eb4bf-e470-4ac2-801e-9150a4afdb5e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.672782 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg" (OuterVolumeSpecName: "kube-api-access-n4bpg") pod "323eb4bf-e470-4ac2-801e-9150a4afdb5e" (UID: "323eb4bf-e470-4ac2-801e-9150a4afdb5e"). InnerVolumeSpecName "kube-api-access-n4bpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769082 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-config\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769164 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7a889f-a301-4b2e-a471-8b481a724509-serving-cert\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769226 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-client-ca\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769349 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfd2t\" (UniqueName: \"kubernetes.io/projected/dd7a889f-a301-4b2e-a471-8b481a724509-kube-api-access-zfd2t\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769445 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmghr\" (UniqueName: \"kubernetes.io/projected/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-kube-api-access-vmghr\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769476 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4bpg\" (UniqueName: \"kubernetes.io/projected/323eb4bf-e470-4ac2-801e-9150a4afdb5e-kube-api-access-n4bpg\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769501 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323eb4bf-e470-4ac2-801e-9150a4afdb5e-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.769525 4836 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.770975 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-config\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.771012 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd7a889f-a301-4b2e-a471-8b481a724509-client-ca\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.776388 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7a889f-a301-4b2e-a471-8b481a724509-serving-cert\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.788224 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfd2t\" (UniqueName: \"kubernetes.io/projected/dd7a889f-a301-4b2e-a471-8b481a724509-kube-api-access-zfd2t\") pod \"route-controller-manager-6997447f6-f5hln\" (UID: \"dd7a889f-a301-4b2e-a471-8b481a724509\") " pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:48:59 crc kubenswrapper[4836]: I0122 09:48:59.882657 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.254929 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln"] Jan 22 09:49:00 crc kubenswrapper[4836]: W0122 09:49:00.262987 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd7a889f_a301_4b2e_a471_8b481a724509.slice/crio-7b535bbb435d324b1edd8738902fedf400769e237d776518a5dc11779715ee61 WatchSource:0}: Error finding container 7b535bbb435d324b1edd8738902fedf400769e237d776518a5dc11779715ee61: Status 404 returned error can't find the container with id 7b535bbb435d324b1edd8738902fedf400769e237d776518a5dc11779715ee61 Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.297747 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" event={"ID":"dd7a889f-a301-4b2e-a471-8b481a724509","Type":"ContainerStarted","Data":"7b535bbb435d324b1edd8738902fedf400769e237d776518a5dc11779715ee61"} Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.299325 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" event={"ID":"323eb4bf-e470-4ac2-801e-9150a4afdb5e","Type":"ContainerDied","Data":"d6fa4580fa06b56760da68adf1a16b7fc08aa7ed0782e6104fe508f9d3ee3c92"} Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.299351 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x" Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.299359 4836 scope.go:117] "RemoveContainer" containerID="91020b0963240054e78b6269ceed248299570bc059636ecb5c0aad0b63014933" Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.301068 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" event={"ID":"0a59c7e5-71fe-4591-ba35-00e94ee8ba6d","Type":"ContainerDied","Data":"ba5c8219c40f16b7817bf5e8a2b73c3bc604bda9988b4fd222df6f5f1d9024f3"} Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.301174 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c45df54bf-gqz5t" Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.316552 4836 scope.go:117] "RemoveContainer" containerID="99b7942769646bafd8d1a4d63f589e249a9d0fe5366c6936afb10bde148141cd" Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.335370 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.338341 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bdf786554-4sv4x"] Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.357556 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:49:00 crc kubenswrapper[4836]: I0122 09:49:00.360588 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c45df54bf-gqz5t"] Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.002819 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" path="/var/lib/kubelet/pods/0a59c7e5-71fe-4591-ba35-00e94ee8ba6d/volumes" Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.003986 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="323eb4bf-e470-4ac2-801e-9150a4afdb5e" path="/var/lib/kubelet/pods/323eb4bf-e470-4ac2-801e-9150a4afdb5e/volumes" Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.312537 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" event={"ID":"dd7a889f-a301-4b2e-a471-8b481a724509","Type":"ContainerStarted","Data":"a8b62e5bb279e8b8dd15594dad5b11915f0371108d53e73af18379b49f92e41f"} Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.312810 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.321932 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" Jan 22 09:49:01 crc kubenswrapper[4836]: I0122 09:49:01.342790 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6997447f6-f5hln" podStartSLOduration=3.3427565169999998 podStartE2EDuration="3.342756517s" podCreationTimestamp="2026-01-22 09:48:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:49:01.33363637 +0000 UTC m=+424.693404917" watchObservedRunningTime="2026-01-22 09:49:01.342756517 +0000 UTC m=+424.702525084" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.123616 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-4mmw8"] Jan 22 09:49:02 crc kubenswrapper[4836]: E0122 09:49:02.124173 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" containerName="controller-manager" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.124193 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" containerName="controller-manager" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.124345 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a59c7e5-71fe-4591-ba35-00e94ee8ba6d" containerName="controller-manager" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.124854 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.128891 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.131392 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.131789 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.131999 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.132001 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.134032 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.137494 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-4mmw8"] Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.140364 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.199231 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-serving-cert\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.199307 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.199344 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-config\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.199364 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-client-ca\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.199402 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd97g\" (UniqueName: \"kubernetes.io/projected/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-kube-api-access-fd97g\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.300865 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-serving-cert\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.300931 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.300965 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-config\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.300984 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-client-ca\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.301012 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd97g\" (UniqueName: \"kubernetes.io/projected/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-kube-api-access-fd97g\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.302311 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-client-ca\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.302423 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-proxy-ca-bundles\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.302854 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-config\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.312499 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-serving-cert\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.322843 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd97g\" (UniqueName: \"kubernetes.io/projected/1eadbfbd-2303-422b-a2a0-bfdb024d14a8-kube-api-access-fd97g\") pod \"controller-manager-685dd7c657-4mmw8\" (UID: \"1eadbfbd-2303-422b-a2a0-bfdb024d14a8\") " pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.441739 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:02 crc kubenswrapper[4836]: I0122 09:49:02.667399 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-685dd7c657-4mmw8"] Jan 22 09:49:03 crc kubenswrapper[4836]: I0122 09:49:03.326458 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" event={"ID":"1eadbfbd-2303-422b-a2a0-bfdb024d14a8","Type":"ContainerStarted","Data":"a99e1a3a730fe7c7c771d76275ec589316686b1537bcf9b5a264c8eae5a0c13f"} Jan 22 09:49:03 crc kubenswrapper[4836]: I0122 09:49:03.326519 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" event={"ID":"1eadbfbd-2303-422b-a2a0-bfdb024d14a8","Type":"ContainerStarted","Data":"e2e7e11b04e4156e6f6685c98c01491529ffd83ce06627e9a402abc4d7c504b5"} Jan 22 09:49:03 crc kubenswrapper[4836]: I0122 09:49:03.342265 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" podStartSLOduration=5.342246567 podStartE2EDuration="5.342246567s" podCreationTimestamp="2026-01-22 09:48:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:49:03.341468597 +0000 UTC m=+426.701237124" watchObservedRunningTime="2026-01-22 09:49:03.342246567 +0000 UTC m=+426.702015084" Jan 22 09:49:04 crc kubenswrapper[4836]: I0122 09:49:04.333581 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:04 crc kubenswrapper[4836]: I0122 09:49:04.338791 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-685dd7c657-4mmw8" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.613368 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mn98s"] Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.614411 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.624444 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mn98s"] Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.700936 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-bound-sa-token\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701001 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/954ebab7-b0d2-4247-91e1-a732527a1206-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701021 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-registry-tls\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701063 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-registry-certificates\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701196 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701241 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-trusted-ca\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701289 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/954ebab7-b0d2-4247-91e1-a732527a1206-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.701333 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlvx4\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-kube-api-access-jlvx4\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.725971 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802238 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlvx4\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-kube-api-access-jlvx4\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802296 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-bound-sa-token\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802334 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/954ebab7-b0d2-4247-91e1-a732527a1206-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802353 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-registry-tls\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802374 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-registry-certificates\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802416 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-trusted-ca\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.802439 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/954ebab7-b0d2-4247-91e1-a732527a1206-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.804440 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/954ebab7-b0d2-4247-91e1-a732527a1206-ca-trust-extracted\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.804976 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-registry-certificates\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.805147 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/954ebab7-b0d2-4247-91e1-a732527a1206-trusted-ca\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.810770 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-registry-tls\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.813429 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/954ebab7-b0d2-4247-91e1-a732527a1206-installation-pull-secrets\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.821977 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlvx4\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-kube-api-access-jlvx4\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.822213 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/954ebab7-b0d2-4247-91e1-a732527a1206-bound-sa-token\") pod \"image-registry-66df7c8f76-mn98s\" (UID: \"954ebab7-b0d2-4247-91e1-a732527a1206\") " pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:08 crc kubenswrapper[4836]: I0122 09:49:08.941561 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:09 crc kubenswrapper[4836]: I0122 09:49:09.415273 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-mn98s"] Jan 22 09:49:10 crc kubenswrapper[4836]: I0122 09:49:10.371158 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" event={"ID":"954ebab7-b0d2-4247-91e1-a732527a1206","Type":"ContainerStarted","Data":"6487493cf28bb80ae943938c24eea67eda159d94beaaa6fb031dce5d937ea779"} Jan 22 09:49:10 crc kubenswrapper[4836]: I0122 09:49:10.371600 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" event={"ID":"954ebab7-b0d2-4247-91e1-a732527a1206","Type":"ContainerStarted","Data":"89b364f504e649dd8eb3b25a6871123064ad106bf8c6210e95a9c3444d1ce5da"} Jan 22 09:49:10 crc kubenswrapper[4836]: I0122 09:49:10.371791 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:10 crc kubenswrapper[4836]: I0122 09:49:10.394405 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" podStartSLOduration=2.394387583 podStartE2EDuration="2.394387583s" podCreationTimestamp="2026-01-22 09:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:49:10.392942116 +0000 UTC m=+433.752710643" watchObservedRunningTime="2026-01-22 09:49:10.394387583 +0000 UTC m=+433.754156120" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.372738 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.379310 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.380069 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7bfkp" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="registry-server" containerID="cri-o://36770dd1b73f0c1834bd32c08cc61f90a2e026f707c90fe90850d12162500a94" gracePeriod=30 Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.381974 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6pq7n" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="registry-server" containerID="cri-o://fb93e0e0be56944808e5cd76ef6d710fbff5cd8eddfca195dc26a0ed9e999695" gracePeriod=30 Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.407480 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.408030 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" containerID="cri-o://a34725fae5b31b25a83b1721699de5384d4639899d3b350b04caeec07e3f1329" gracePeriod=30 Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.414722 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.415177 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ztdcc" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="registry-server" containerID="cri-o://7414f3234b7b8f2b7cf7c340e91226d1f57f742820b1eddb0c7d52fe22d2942f" gracePeriod=30 Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.419125 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.419498 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cpx6h" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="registry-server" containerID="cri-o://48c3d6f1bc3a105595acf43ce9dd9a638cda668ec974014d73694c7cc3a0db1a" gracePeriod=30 Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.423408 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crc7k"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.424275 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.449472 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crc7k"] Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.542300 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.542352 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkgz6\" (UniqueName: \"kubernetes.io/projected/66077fb8-1811-4d25-8b81-9ca672521135-kube-api-access-nkgz6\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.542371 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.643926 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.645156 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.645302 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkgz6\" (UniqueName: \"kubernetes.io/projected/66077fb8-1811-4d25-8b81-9ca672521135-kube-api-access-nkgz6\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.645430 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.651662 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/66077fb8-1811-4d25-8b81-9ca672521135-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.674832 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkgz6\" (UniqueName: \"kubernetes.io/projected/66077fb8-1811-4d25-8b81-9ca672521135-kube-api-access-nkgz6\") pod \"marketplace-operator-79b997595-crc7k\" (UID: \"66077fb8-1811-4d25-8b81-9ca672521135\") " pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:11 crc kubenswrapper[4836]: I0122 09:49:11.746670 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.154513 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-crc7k"] Jan 22 09:49:12 crc kubenswrapper[4836]: W0122 09:49:12.191715 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66077fb8_1811_4d25_8b81_9ca672521135.slice/crio-43a8e9358d1b251d79a6a31e39353bfe106e0bcc354139e1c7110ea515dbfec2 WatchSource:0}: Error finding container 43a8e9358d1b251d79a6a31e39353bfe106e0bcc354139e1c7110ea515dbfec2: Status 404 returned error can't find the container with id 43a8e9358d1b251d79a6a31e39353bfe106e0bcc354139e1c7110ea515dbfec2 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.387851 4836 generic.go:334] "Generic (PLEG): container finished" podID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerID="7414f3234b7b8f2b7cf7c340e91226d1f57f742820b1eddb0c7d52fe22d2942f" exitCode=0 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.387932 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerDied","Data":"7414f3234b7b8f2b7cf7c340e91226d1f57f742820b1eddb0c7d52fe22d2942f"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.390796 4836 generic.go:334] "Generic (PLEG): container finished" podID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerID="36770dd1b73f0c1834bd32c08cc61f90a2e026f707c90fe90850d12162500a94" exitCode=0 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.390850 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerDied","Data":"36770dd1b73f0c1834bd32c08cc61f90a2e026f707c90fe90850d12162500a94"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.392736 4836 generic.go:334] "Generic (PLEG): container finished" podID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerID="fb93e0e0be56944808e5cd76ef6d710fbff5cd8eddfca195dc26a0ed9e999695" exitCode=0 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.392770 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerDied","Data":"fb93e0e0be56944808e5cd76ef6d710fbff5cd8eddfca195dc26a0ed9e999695"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.394801 4836 generic.go:334] "Generic (PLEG): container finished" podID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerID="48c3d6f1bc3a105595acf43ce9dd9a638cda668ec974014d73694c7cc3a0db1a" exitCode=0 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.394855 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerDied","Data":"48c3d6f1bc3a105595acf43ce9dd9a638cda668ec974014d73694c7cc3a0db1a"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.396596 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" event={"ID":"66077fb8-1811-4d25-8b81-9ca672521135","Type":"ContainerStarted","Data":"709c35464733f40da75fdeab5f044e80501ac80ce6bc228512e5e09242b94bff"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.396620 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" event={"ID":"66077fb8-1811-4d25-8b81-9ca672521135","Type":"ContainerStarted","Data":"43a8e9358d1b251d79a6a31e39353bfe106e0bcc354139e1c7110ea515dbfec2"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.397428 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.398918 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4bdcx_ee9368ed-7945-471a-9869-aa2fdb8ad2c7/marketplace-operator/1.log" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.398960 4836 generic.go:334] "Generic (PLEG): container finished" podID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerID="a34725fae5b31b25a83b1721699de5384d4639899d3b350b04caeec07e3f1329" exitCode=0 Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.398992 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerDied","Data":"a34725fae5b31b25a83b1721699de5384d4639899d3b350b04caeec07e3f1329"} Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.399025 4836 scope.go:117] "RemoveContainer" containerID="8e76884df83b57793e7a8bb7cc74efe13deb0b1135f5726fb12697e1d89866a2" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.399180 4836 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-crc7k container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" start-of-body= Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.399215 4836 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" podUID="66077fb8-1811-4d25-8b81-9ca672521135" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.71:8080/healthz\": dial tcp 10.217.0.71:8080: connect: connection refused" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.413489 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" podStartSLOduration=1.4134692420000001 podStartE2EDuration="1.413469242s" podCreationTimestamp="2026-01-22 09:49:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:49:12.412913419 +0000 UTC m=+435.772681936" watchObservedRunningTime="2026-01-22 09:49:12.413469242 +0000 UTC m=+435.773237759" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.550581 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.661437 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics\") pod \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.661909 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n89t5\" (UniqueName: \"kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5\") pod \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.662010 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca\") pod \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\" (UID: \"ee9368ed-7945-471a-9869-aa2fdb8ad2c7\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.662520 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ee9368ed-7945-471a-9869-aa2fdb8ad2c7" (UID: "ee9368ed-7945-471a-9869-aa2fdb8ad2c7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.670530 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5" (OuterVolumeSpecName: "kube-api-access-n89t5") pod "ee9368ed-7945-471a-9869-aa2fdb8ad2c7" (UID: "ee9368ed-7945-471a-9869-aa2fdb8ad2c7"). InnerVolumeSpecName "kube-api-access-n89t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.677030 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ee9368ed-7945-471a-9869-aa2fdb8ad2c7" (UID: "ee9368ed-7945-471a-9869-aa2fdb8ad2c7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.741713 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.747950 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.752726 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.763125 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n89t5\" (UniqueName: \"kubernetes.io/projected/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-kube-api-access-n89t5\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.763155 4836 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.763164 4836 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee9368ed-7945-471a-9869-aa2fdb8ad2c7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.811456 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864425 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities\") pod \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864474 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crrh\" (UniqueName: \"kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh\") pod \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864501 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities\") pod \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864571 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content\") pod \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864586 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9m69\" (UniqueName: \"kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69\") pod \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864615 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities\") pod \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864641 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content\") pod \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\" (UID: \"f89c88a4-5375-47f6-ab66-07aa49a11a2e\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864665 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content\") pod \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\" (UID: \"d7de3ebe-e601-46b3-bda0-8e08d0a42e43\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.864697 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfr6h\" (UniqueName: \"kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h\") pod \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\" (UID: \"65b61041-36cc-4c4d-be61-dd4e370b7ff3\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.865686 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities" (OuterVolumeSpecName: "utilities") pod "65b61041-36cc-4c4d-be61-dd4e370b7ff3" (UID: "65b61041-36cc-4c4d-be61-dd4e370b7ff3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.866270 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities" (OuterVolumeSpecName: "utilities") pod "f89c88a4-5375-47f6-ab66-07aa49a11a2e" (UID: "f89c88a4-5375-47f6-ab66-07aa49a11a2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.866398 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities" (OuterVolumeSpecName: "utilities") pod "d7de3ebe-e601-46b3-bda0-8e08d0a42e43" (UID: "d7de3ebe-e601-46b3-bda0-8e08d0a42e43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.867848 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh" (OuterVolumeSpecName: "kube-api-access-2crrh") pod "d7de3ebe-e601-46b3-bda0-8e08d0a42e43" (UID: "d7de3ebe-e601-46b3-bda0-8e08d0a42e43"). InnerVolumeSpecName "kube-api-access-2crrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.868086 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h" (OuterVolumeSpecName: "kube-api-access-wfr6h") pod "65b61041-36cc-4c4d-be61-dd4e370b7ff3" (UID: "65b61041-36cc-4c4d-be61-dd4e370b7ff3"). InnerVolumeSpecName "kube-api-access-wfr6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.868721 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69" (OuterVolumeSpecName: "kube-api-access-j9m69") pod "f89c88a4-5375-47f6-ab66-07aa49a11a2e" (UID: "f89c88a4-5375-47f6-ab66-07aa49a11a2e"). InnerVolumeSpecName "kube-api-access-j9m69". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.892754 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f89c88a4-5375-47f6-ab66-07aa49a11a2e" (UID: "f89c88a4-5375-47f6-ab66-07aa49a11a2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.911450 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7de3ebe-e601-46b3-bda0-8e08d0a42e43" (UID: "d7de3ebe-e601-46b3-bda0-8e08d0a42e43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.922812 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65b61041-36cc-4c4d-be61-dd4e370b7ff3" (UID: "65b61041-36cc-4c4d-be61-dd4e370b7ff3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966291 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jws8b\" (UniqueName: \"kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b\") pod \"4363607a-4a5d-4e80-9389-4cd4214e316b\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966367 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content\") pod \"4363607a-4a5d-4e80-9389-4cd4214e316b\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966389 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities\") pod \"4363607a-4a5d-4e80-9389-4cd4214e316b\" (UID: \"4363607a-4a5d-4e80-9389-4cd4214e316b\") " Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966662 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966680 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89c88a4-5375-47f6-ab66-07aa49a11a2e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966689 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966697 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfr6h\" (UniqueName: \"kubernetes.io/projected/65b61041-36cc-4c4d-be61-dd4e370b7ff3-kube-api-access-wfr6h\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966706 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966713 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crrh\" (UniqueName: \"kubernetes.io/projected/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-kube-api-access-2crrh\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966722 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7de3ebe-e601-46b3-bda0-8e08d0a42e43-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966730 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65b61041-36cc-4c4d-be61-dd4e370b7ff3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.966739 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9m69\" (UniqueName: \"kubernetes.io/projected/f89c88a4-5375-47f6-ab66-07aa49a11a2e-kube-api-access-j9m69\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.967372 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities" (OuterVolumeSpecName: "utilities") pod "4363607a-4a5d-4e80-9389-4cd4214e316b" (UID: "4363607a-4a5d-4e80-9389-4cd4214e316b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:12 crc kubenswrapper[4836]: I0122 09:49:12.968942 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b" (OuterVolumeSpecName: "kube-api-access-jws8b") pod "4363607a-4a5d-4e80-9389-4cd4214e316b" (UID: "4363607a-4a5d-4e80-9389-4cd4214e316b"). InnerVolumeSpecName "kube-api-access-jws8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.068178 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jws8b\" (UniqueName: \"kubernetes.io/projected/4363607a-4a5d-4e80-9389-4cd4214e316b-kube-api-access-jws8b\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.068547 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.082657 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4363607a-4a5d-4e80-9389-4cd4214e316b" (UID: "4363607a-4a5d-4e80-9389-4cd4214e316b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.169748 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4363607a-4a5d-4e80-9389-4cd4214e316b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.405814 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" event={"ID":"ee9368ed-7945-471a-9869-aa2fdb8ad2c7","Type":"ContainerDied","Data":"d19eefba4f8fce67a7636a84a2b27458f3e34dc75e2855e05caad53d0af07f9f"} Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.405868 4836 scope.go:117] "RemoveContainer" containerID="a34725fae5b31b25a83b1721699de5384d4639899d3b350b04caeec07e3f1329" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.406182 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4bdcx" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.414172 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztdcc" event={"ID":"f89c88a4-5375-47f6-ab66-07aa49a11a2e","Type":"ContainerDied","Data":"c696a1844cb19bd078e0ee2e9c34d50c9f419109328e5a8bffa2cfaa5da39653"} Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.414243 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztdcc" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.424469 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bfkp" event={"ID":"65b61041-36cc-4c4d-be61-dd4e370b7ff3","Type":"ContainerDied","Data":"2eb611cec8de73c43e2cafcd79b7b9256db285b8f3fa1e288977f3002f349767"} Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.424488 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bfkp" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.429342 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.429695 4836 scope.go:117] "RemoveContainer" containerID="7414f3234b7b8f2b7cf7c340e91226d1f57f742820b1eddb0c7d52fe22d2942f" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.429886 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pq7n" event={"ID":"d7de3ebe-e601-46b3-bda0-8e08d0a42e43","Type":"ContainerDied","Data":"282575cb2ba5fcd5ed74d9f5b2755e7eb0b92cfb5fe2253e3bc58f88ba366c61"} Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.429911 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pq7n" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.434052 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cpx6h" event={"ID":"4363607a-4a5d-4e80-9389-4cd4214e316b","Type":"ContainerDied","Data":"55be1a33fa4ebccbe9e46fdb32f4de6b6136879f08315b21a4e95eeb34d1997e"} Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.434139 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cpx6h" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.435280 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4bdcx"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.436098 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-crc7k" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.446575 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.453806 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztdcc"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.461887 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.462175 4836 scope.go:117] "RemoveContainer" containerID="da71b4d14abfbe08e3081ff13ceb0bc2c9bee41727ad2596a91aec994053679c" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.468661 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7bfkp"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.473933 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.484023 4836 scope.go:117] "RemoveContainer" containerID="aebb3b8fd635bcbc64d2ba7a6d6727c22de7245cb4231d8e8104b4ff6be8a4f3" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.485756 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6pq7n"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.517730 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.521617 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cpx6h"] Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.534788 4836 scope.go:117] "RemoveContainer" containerID="36770dd1b73f0c1834bd32c08cc61f90a2e026f707c90fe90850d12162500a94" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.556536 4836 scope.go:117] "RemoveContainer" containerID="a2644ef719bbefec72d05031434fac6410e5e6f6f0f0241229bd289c03a41053" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.571966 4836 scope.go:117] "RemoveContainer" containerID="7a1fb1404f277b34aceb3fd88bd162f157327df6be57dcde84a7a2504c241b10" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.585830 4836 scope.go:117] "RemoveContainer" containerID="fb93e0e0be56944808e5cd76ef6d710fbff5cd8eddfca195dc26a0ed9e999695" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.604587 4836 scope.go:117] "RemoveContainer" containerID="cab067f8b7d4b0b584700bf162084895d67a82366e85cd69d45665c958f885d0" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.618441 4836 scope.go:117] "RemoveContainer" containerID="6d3166cd0b03f93eef234b51531da6330dcaac837a67db0bb4acd8e6762a9a6d" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.641324 4836 scope.go:117] "RemoveContainer" containerID="48c3d6f1bc3a105595acf43ce9dd9a638cda668ec974014d73694c7cc3a0db1a" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.652166 4836 scope.go:117] "RemoveContainer" containerID="83bbbbbfea658121a9494c11abdda321e1e6a6af25b144e68c4dc6c32e2f327b" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.666529 4836 scope.go:117] "RemoveContainer" containerID="15bd7ceb1ed697fafeddee904695f4d1b182a3c54c84cffac30645dae1b0fbff" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991385 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m67hn"] Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991597 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991610 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991620 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991627 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991635 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991641 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991649 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991654 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991662 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991667 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991676 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991681 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991689 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991694 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991704 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991709 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991718 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991725 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991734 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991743 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991752 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991759 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991773 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991779 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991785 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991791 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="extract-content" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.991798 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991803 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="extract-utilities" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991883 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991893 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991902 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991912 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991921 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.991929 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" containerName="registry-server" Jan 22 09:49:13 crc kubenswrapper[4836]: E0122 09:49:13.992013 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.992019 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.992108 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" containerName="marketplace-operator" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.992637 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:13 crc kubenswrapper[4836]: I0122 09:49:13.994717 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.000332 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m67hn"] Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.081332 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-catalog-content\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.081603 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7chj5\" (UniqueName: \"kubernetes.io/projected/d131c247-f421-4d22-8e20-4605e2e6cd18-kube-api-access-7chj5\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.081809 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-utilities\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.183565 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-utilities\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.183833 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-catalog-content\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.183950 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7chj5\" (UniqueName: \"kubernetes.io/projected/d131c247-f421-4d22-8e20-4605e2e6cd18-kube-api-access-7chj5\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.184783 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-utilities\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.184825 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d131c247-f421-4d22-8e20-4605e2e6cd18-catalog-content\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.203364 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7chj5\" (UniqueName: \"kubernetes.io/projected/d131c247-f421-4d22-8e20-4605e2e6cd18-kube-api-access-7chj5\") pod \"redhat-marketplace-m67hn\" (UID: \"d131c247-f421-4d22-8e20-4605e2e6cd18\") " pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.309271 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:14 crc kubenswrapper[4836]: I0122 09:49:14.694666 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m67hn"] Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.003439 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4363607a-4a5d-4e80-9389-4cd4214e316b" path="/var/lib/kubelet/pods/4363607a-4a5d-4e80-9389-4cd4214e316b/volumes" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.004419 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b61041-36cc-4c4d-be61-dd4e370b7ff3" path="/var/lib/kubelet/pods/65b61041-36cc-4c4d-be61-dd4e370b7ff3/volumes" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.005181 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7de3ebe-e601-46b3-bda0-8e08d0a42e43" path="/var/lib/kubelet/pods/d7de3ebe-e601-46b3-bda0-8e08d0a42e43/volumes" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.006427 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9368ed-7945-471a-9869-aa2fdb8ad2c7" path="/var/lib/kubelet/pods/ee9368ed-7945-471a-9869-aa2fdb8ad2c7/volumes" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.007026 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89c88a4-5375-47f6-ab66-07aa49a11a2e" path="/var/lib/kubelet/pods/f89c88a4-5375-47f6-ab66-07aa49a11a2e/volumes" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.385953 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zrlz2"] Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.387192 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.389003 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.396249 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrlz2"] Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.452248 4836 generic.go:334] "Generic (PLEG): container finished" podID="d131c247-f421-4d22-8e20-4605e2e6cd18" containerID="56e404f48889b8490d15dc70141a2402a9c363b2adfb59f9c0afd620c189ed70" exitCode=0 Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.453247 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m67hn" event={"ID":"d131c247-f421-4d22-8e20-4605e2e6cd18","Type":"ContainerDied","Data":"56e404f48889b8490d15dc70141a2402a9c363b2adfb59f9c0afd620c189ed70"} Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.453686 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m67hn" event={"ID":"d131c247-f421-4d22-8e20-4605e2e6cd18","Type":"ContainerStarted","Data":"586a555f96a6598f1276e0e6174cff75c21cd56d8307392cc37f0d01a72792d3"} Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.504834 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-utilities\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.504895 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-catalog-content\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.504947 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxjhw\" (UniqueName: \"kubernetes.io/projected/a93e2393-3439-4b33-ac03-142faa9de7de-kube-api-access-pxjhw\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.605690 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-utilities\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.605804 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-catalog-content\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.605843 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxjhw\" (UniqueName: \"kubernetes.io/projected/a93e2393-3439-4b33-ac03-142faa9de7de-kube-api-access-pxjhw\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.606205 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-utilities\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.606368 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a93e2393-3439-4b33-ac03-142faa9de7de-catalog-content\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.637902 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxjhw\" (UniqueName: \"kubernetes.io/projected/a93e2393-3439-4b33-ac03-142faa9de7de-kube-api-access-pxjhw\") pod \"redhat-operators-zrlz2\" (UID: \"a93e2393-3439-4b33-ac03-142faa9de7de\") " pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:15 crc kubenswrapper[4836]: I0122 09:49:15.702433 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.132415 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrlz2"] Jan 22 09:49:16 crc kubenswrapper[4836]: W0122 09:49:16.152673 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda93e2393_3439_4b33_ac03_142faa9de7de.slice/crio-5fc8388da714a6579b418733c917731b4b1356b414fd63bd3e8018ab7bdf035a WatchSource:0}: Error finding container 5fc8388da714a6579b418733c917731b4b1356b414fd63bd3e8018ab7bdf035a: Status 404 returned error can't find the container with id 5fc8388da714a6579b418733c917731b4b1356b414fd63bd3e8018ab7bdf035a Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.383358 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z48bp"] Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.384340 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.387544 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.404098 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z48bp"] Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.458798 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrlz2" event={"ID":"a93e2393-3439-4b33-ac03-142faa9de7de","Type":"ContainerStarted","Data":"5fc8388da714a6579b418733c917731b4b1356b414fd63bd3e8018ab7bdf035a"} Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.516575 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jftg\" (UniqueName: \"kubernetes.io/projected/8d7f3c8f-bc53-4fb9-948a-408169582191-kube-api-access-4jftg\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.516734 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-utilities\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.516758 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-catalog-content\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.618075 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jftg\" (UniqueName: \"kubernetes.io/projected/8d7f3c8f-bc53-4fb9-948a-408169582191-kube-api-access-4jftg\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.618146 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-utilities\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.618172 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-catalog-content\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.618618 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-catalog-content\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.618768 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d7f3c8f-bc53-4fb9-948a-408169582191-utilities\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.637391 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jftg\" (UniqueName: \"kubernetes.io/projected/8d7f3c8f-bc53-4fb9-948a-408169582191-kube-api-access-4jftg\") pod \"community-operators-z48bp\" (UID: \"8d7f3c8f-bc53-4fb9-948a-408169582191\") " pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:16 crc kubenswrapper[4836]: I0122 09:49:16.727327 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.788967 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.790304 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.792161 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.800514 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.959736 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.959821 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:17 crc kubenswrapper[4836]: I0122 09:49:17.959857 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.062528 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.062638 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.062707 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.063935 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.063972 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.088967 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp\") pod \"certified-operators-wxfgn\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.109029 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.332672 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z48bp"] Jan 22 09:49:18 crc kubenswrapper[4836]: W0122 09:49:18.345294 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d7f3c8f_bc53_4fb9_948a_408169582191.slice/crio-57bd64f99c631eb4a99ca810767fd2d761c2396ee6138f0de6df58c664f4b3fc WatchSource:0}: Error finding container 57bd64f99c631eb4a99ca810767fd2d761c2396ee6138f0de6df58c664f4b3fc: Status 404 returned error can't find the container with id 57bd64f99c631eb4a99ca810767fd2d761c2396ee6138f0de6df58c664f4b3fc Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.470628 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z48bp" event={"ID":"8d7f3c8f-bc53-4fb9-948a-408169582191","Type":"ContainerStarted","Data":"f5774da25b6bcc9e5e292bd022e348f0e2d67b05d5cf395c87f23ef4f14b094f"} Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.470672 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z48bp" event={"ID":"8d7f3c8f-bc53-4fb9-948a-408169582191","Type":"ContainerStarted","Data":"57bd64f99c631eb4a99ca810767fd2d761c2396ee6138f0de6df58c664f4b3fc"} Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.472398 4836 generic.go:334] "Generic (PLEG): container finished" podID="a93e2393-3439-4b33-ac03-142faa9de7de" containerID="b0e0af982b34e861ebb37855ab551f20f3d20b67e566c1e3ff4998c77e87fa87" exitCode=0 Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.472787 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrlz2" event={"ID":"a93e2393-3439-4b33-ac03-142faa9de7de","Type":"ContainerDied","Data":"b0e0af982b34e861ebb37855ab551f20f3d20b67e566c1e3ff4998c77e87fa87"} Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.475206 4836 generic.go:334] "Generic (PLEG): container finished" podID="d131c247-f421-4d22-8e20-4605e2e6cd18" containerID="21998ce70beb227bbd414df50334d680940cd8379be4457f15e6035309dfeff5" exitCode=0 Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.475254 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m67hn" event={"ID":"d131c247-f421-4d22-8e20-4605e2e6cd18","Type":"ContainerDied","Data":"21998ce70beb227bbd414df50334d680940cd8379be4457f15e6035309dfeff5"} Jan 22 09:49:18 crc kubenswrapper[4836]: I0122 09:49:18.507477 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:49:18 crc kubenswrapper[4836]: W0122 09:49:18.564380 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb79caeee_9bca_44cf_89ac_520e9b0732cc.slice/crio-2bf91bd1710e71b2f385c64dd3be3911589e45424657645c49635556cd70b59d WatchSource:0}: Error finding container 2bf91bd1710e71b2f385c64dd3be3911589e45424657645c49635556cd70b59d: Status 404 returned error can't find the container with id 2bf91bd1710e71b2f385c64dd3be3911589e45424657645c49635556cd70b59d Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.492489 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrlz2" event={"ID":"a93e2393-3439-4b33-ac03-142faa9de7de","Type":"ContainerStarted","Data":"33bc8b4b4c4c0fbbaf7266c6f8207652fcbdbfcc1d864afdb50e97d0245c277c"} Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.500512 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m67hn" event={"ID":"d131c247-f421-4d22-8e20-4605e2e6cd18","Type":"ContainerStarted","Data":"c9606f46f4c52da1f85b17ae19c080e53277eb7984beed57cfd40e729bc80b8c"} Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.501787 4836 generic.go:334] "Generic (PLEG): container finished" podID="8d7f3c8f-bc53-4fb9-948a-408169582191" containerID="f5774da25b6bcc9e5e292bd022e348f0e2d67b05d5cf395c87f23ef4f14b094f" exitCode=0 Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.501842 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z48bp" event={"ID":"8d7f3c8f-bc53-4fb9-948a-408169582191","Type":"ContainerDied","Data":"f5774da25b6bcc9e5e292bd022e348f0e2d67b05d5cf395c87f23ef4f14b094f"} Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.504101 4836 generic.go:334] "Generic (PLEG): container finished" podID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerID="2ee5df5f7b69fcd7e663472bedd6b7f581de1db615ad2493e0bdd1ccbf522342" exitCode=0 Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.504133 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerDied","Data":"2ee5df5f7b69fcd7e663472bedd6b7f581de1db615ad2493e0bdd1ccbf522342"} Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.504182 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerStarted","Data":"2bf91bd1710e71b2f385c64dd3be3911589e45424657645c49635556cd70b59d"} Jan 22 09:49:19 crc kubenswrapper[4836]: I0122 09:49:19.530334 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m67hn" podStartSLOduration=3.080216266 podStartE2EDuration="6.530312885s" podCreationTimestamp="2026-01-22 09:49:13 +0000 UTC" firstStartedPulling="2026-01-22 09:49:15.454746062 +0000 UTC m=+438.814514609" lastFinishedPulling="2026-01-22 09:49:18.904842711 +0000 UTC m=+442.264611228" observedRunningTime="2026-01-22 09:49:19.527184767 +0000 UTC m=+442.886953284" watchObservedRunningTime="2026-01-22 09:49:19.530312885 +0000 UTC m=+442.890081402" Jan 22 09:49:21 crc kubenswrapper[4836]: I0122 09:49:21.516818 4836 generic.go:334] "Generic (PLEG): container finished" podID="a93e2393-3439-4b33-ac03-142faa9de7de" containerID="33bc8b4b4c4c0fbbaf7266c6f8207652fcbdbfcc1d864afdb50e97d0245c277c" exitCode=0 Jan 22 09:49:21 crc kubenswrapper[4836]: I0122 09:49:21.516895 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrlz2" event={"ID":"a93e2393-3439-4b33-ac03-142faa9de7de","Type":"ContainerDied","Data":"33bc8b4b4c4c0fbbaf7266c6f8207652fcbdbfcc1d864afdb50e97d0245c277c"} Jan 22 09:49:24 crc kubenswrapper[4836]: I0122 09:49:24.310398 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:24 crc kubenswrapper[4836]: I0122 09:49:24.310957 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:24 crc kubenswrapper[4836]: I0122 09:49:24.347691 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:24 crc kubenswrapper[4836]: I0122 09:49:24.581185 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m67hn" Jan 22 09:49:25 crc kubenswrapper[4836]: I0122 09:49:25.545586 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrlz2" event={"ID":"a93e2393-3439-4b33-ac03-142faa9de7de","Type":"ContainerStarted","Data":"cca1b7cedd3bcb0e526e60f24e921343655cabcf0f03b49cfcbe8ff19af6b878"} Jan 22 09:49:25 crc kubenswrapper[4836]: I0122 09:49:25.547723 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerStarted","Data":"2f82c9addd08a4cbdb2162c8655b28fefed9649efa2fd4c6a6447aa90ef42385"} Jan 22 09:49:26 crc kubenswrapper[4836]: I0122 09:49:26.554554 4836 generic.go:334] "Generic (PLEG): container finished" podID="8d7f3c8f-bc53-4fb9-948a-408169582191" containerID="a90ca39e3aebc6781b3b4162b59590fb4e2964929ac5c975d8deba4c3ade5b56" exitCode=0 Jan 22 09:49:26 crc kubenswrapper[4836]: I0122 09:49:26.554644 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z48bp" event={"ID":"8d7f3c8f-bc53-4fb9-948a-408169582191","Type":"ContainerDied","Data":"a90ca39e3aebc6781b3b4162b59590fb4e2964929ac5c975d8deba4c3ade5b56"} Jan 22 09:49:26 crc kubenswrapper[4836]: I0122 09:49:26.558327 4836 generic.go:334] "Generic (PLEG): container finished" podID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerID="2f82c9addd08a4cbdb2162c8655b28fefed9649efa2fd4c6a6447aa90ef42385" exitCode=0 Jan 22 09:49:26 crc kubenswrapper[4836]: I0122 09:49:26.558741 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerDied","Data":"2f82c9addd08a4cbdb2162c8655b28fefed9649efa2fd4c6a6447aa90ef42385"} Jan 22 09:49:26 crc kubenswrapper[4836]: I0122 09:49:26.594990 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zrlz2" podStartSLOduration=6.832040361 podStartE2EDuration="11.594972435s" podCreationTimestamp="2026-01-22 09:49:15 +0000 UTC" firstStartedPulling="2026-01-22 09:49:18.473707785 +0000 UTC m=+441.833476302" lastFinishedPulling="2026-01-22 09:49:23.236639859 +0000 UTC m=+446.596408376" observedRunningTime="2026-01-22 09:49:26.594867982 +0000 UTC m=+449.954636509" watchObservedRunningTime="2026-01-22 09:49:26.594972435 +0000 UTC m=+449.954740952" Jan 22 09:49:28 crc kubenswrapper[4836]: I0122 09:49:28.573112 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z48bp" event={"ID":"8d7f3c8f-bc53-4fb9-948a-408169582191","Type":"ContainerStarted","Data":"97e4d8a818027c3583a916b16bdd076c91cd8f04e897c46c2ade3d4b2a98e440"} Jan 22 09:49:28 crc kubenswrapper[4836]: I0122 09:49:28.576875 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerStarted","Data":"c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91"} Jan 22 09:49:28 crc kubenswrapper[4836]: I0122 09:49:28.596076 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z48bp" podStartSLOduration=4.753711555 podStartE2EDuration="12.596038993s" podCreationTimestamp="2026-01-22 09:49:16 +0000 UTC" firstStartedPulling="2026-01-22 09:49:19.503122516 +0000 UTC m=+442.862891033" lastFinishedPulling="2026-01-22 09:49:27.345449954 +0000 UTC m=+450.705218471" observedRunningTime="2026-01-22 09:49:28.592226118 +0000 UTC m=+451.951994636" watchObservedRunningTime="2026-01-22 09:49:28.596038993 +0000 UTC m=+451.955807510" Jan 22 09:49:28 crc kubenswrapper[4836]: I0122 09:49:28.610124 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wxfgn" podStartSLOduration=3.622047114 podStartE2EDuration="11.610101425s" podCreationTimestamp="2026-01-22 09:49:17 +0000 UTC" firstStartedPulling="2026-01-22 09:49:19.505879114 +0000 UTC m=+442.865647631" lastFinishedPulling="2026-01-22 09:49:27.493933425 +0000 UTC m=+450.853701942" observedRunningTime="2026-01-22 09:49:28.60668201 +0000 UTC m=+451.966450527" watchObservedRunningTime="2026-01-22 09:49:28.610101425 +0000 UTC m=+451.969869942" Jan 22 09:49:28 crc kubenswrapper[4836]: I0122 09:49:28.954701 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-mn98s" Jan 22 09:49:29 crc kubenswrapper[4836]: I0122 09:49:29.018386 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:49:35 crc kubenswrapper[4836]: I0122 09:49:35.703570 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:35 crc kubenswrapper[4836]: I0122 09:49:35.704153 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:35 crc kubenswrapper[4836]: I0122 09:49:35.740845 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:36 crc kubenswrapper[4836]: I0122 09:49:36.663812 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zrlz2" Jan 22 09:49:36 crc kubenswrapper[4836]: I0122 09:49:36.728673 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:36 crc kubenswrapper[4836]: I0122 09:49:36.729000 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:36 crc kubenswrapper[4836]: I0122 09:49:36.768875 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:37 crc kubenswrapper[4836]: I0122 09:49:37.680197 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z48bp" Jan 22 09:49:38 crc kubenswrapper[4836]: I0122 09:49:38.109440 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:38 crc kubenswrapper[4836]: I0122 09:49:38.109478 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:38 crc kubenswrapper[4836]: I0122 09:49:38.170539 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:38 crc kubenswrapper[4836]: I0122 09:49:38.673021 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:49:54 crc kubenswrapper[4836]: I0122 09:49:54.060340 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" podUID="f60d64f9-1517-4093-9a99-abd333d12a36" containerName="registry" containerID="cri-o://f519e9ae5d8ad9daa7a64f7bf62caa091a82282096d19c04e717ec50a178f20c" gracePeriod=30 Jan 22 09:49:54 crc kubenswrapper[4836]: I0122 09:49:54.742926 4836 generic.go:334] "Generic (PLEG): container finished" podID="f60d64f9-1517-4093-9a99-abd333d12a36" containerID="f519e9ae5d8ad9daa7a64f7bf62caa091a82282096d19c04e717ec50a178f20c" exitCode=0 Jan 22 09:49:54 crc kubenswrapper[4836]: I0122 09:49:54.742967 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" event={"ID":"f60d64f9-1517-4093-9a99-abd333d12a36","Type":"ContainerDied","Data":"f519e9ae5d8ad9daa7a64f7bf62caa091a82282096d19c04e717ec50a178f20c"} Jan 22 09:49:54 crc kubenswrapper[4836]: I0122 09:49:54.988685 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060358 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060407 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krspr\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060550 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060577 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060597 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060622 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060638 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.060665 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted\") pod \"f60d64f9-1517-4093-9a99-abd333d12a36\" (UID: \"f60d64f9-1517-4093-9a99-abd333d12a36\") " Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.061691 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.061915 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.066270 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr" (OuterVolumeSpecName: "kube-api-access-krspr") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "kube-api-access-krspr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.066288 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.066504 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.066700 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.070863 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.075339 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f60d64f9-1517-4093-9a99-abd333d12a36" (UID: "f60d64f9-1517-4093-9a99-abd333d12a36"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161488 4836 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161526 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krspr\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-kube-api-access-krspr\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161537 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161546 4836 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f60d64f9-1517-4093-9a99-abd333d12a36-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161554 4836 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f60d64f9-1517-4093-9a99-abd333d12a36-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161562 4836 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f60d64f9-1517-4093-9a99-abd333d12a36-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.161570 4836 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f60d64f9-1517-4093-9a99-abd333d12a36-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.753279 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" event={"ID":"f60d64f9-1517-4093-9a99-abd333d12a36","Type":"ContainerDied","Data":"59e7e33d832f61a2e3f267769ce208f515c2fad8fff6af6b97cccd7206e3944a"} Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.753378 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t4zmg" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.753381 4836 scope.go:117] "RemoveContainer" containerID="f519e9ae5d8ad9daa7a64f7bf62caa091a82282096d19c04e717ec50a178f20c" Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.813773 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:49:55 crc kubenswrapper[4836]: I0122 09:49:55.820804 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t4zmg"] Jan 22 09:49:57 crc kubenswrapper[4836]: I0122 09:49:57.012675 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f60d64f9-1517-4093-9a99-abd333d12a36" path="/var/lib/kubelet/pods/f60d64f9-1517-4093-9a99-abd333d12a36/volumes" Jan 22 09:51:23 crc kubenswrapper[4836]: I0122 09:51:23.897405 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:51:23 crc kubenswrapper[4836]: I0122 09:51:23.898019 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:51:53 crc kubenswrapper[4836]: I0122 09:51:53.897900 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:51:53 crc kubenswrapper[4836]: I0122 09:51:53.898585 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:52:23 crc kubenswrapper[4836]: I0122 09:52:23.897646 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:52:23 crc kubenswrapper[4836]: I0122 09:52:23.898230 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:52:23 crc kubenswrapper[4836]: I0122 09:52:23.898293 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:52:23 crc kubenswrapper[4836]: I0122 09:52:23.899087 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 09:52:23 crc kubenswrapper[4836]: I0122 09:52:23.899206 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401" gracePeriod=600 Jan 22 09:52:24 crc kubenswrapper[4836]: I0122 09:52:24.629283 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401" exitCode=0 Jan 22 09:52:24 crc kubenswrapper[4836]: I0122 09:52:24.629344 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401"} Jan 22 09:52:24 crc kubenswrapper[4836]: I0122 09:52:24.629624 4836 scope.go:117] "RemoveContainer" containerID="5c3da6bc0e8e4b94239280120ea03fd03e99ba9f1aef60f8f6bd740963ac9f86" Jan 22 09:52:25 crc kubenswrapper[4836]: I0122 09:52:25.640931 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd"} Jan 22 09:54:29 crc kubenswrapper[4836]: I0122 09:54:29.427292 4836 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 22 09:54:53 crc kubenswrapper[4836]: I0122 09:54:53.897396 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:54:53 crc kubenswrapper[4836]: I0122 09:54:53.897892 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:55:23 crc kubenswrapper[4836]: I0122 09:55:23.897636 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:55:23 crc kubenswrapper[4836]: I0122 09:55:23.898622 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:55:53 crc kubenswrapper[4836]: I0122 09:55:53.897715 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:55:53 crc kubenswrapper[4836]: I0122 09:55:53.898397 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:55:53 crc kubenswrapper[4836]: I0122 09:55:53.898467 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:55:53 crc kubenswrapper[4836]: I0122 09:55:53.899448 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 09:55:53 crc kubenswrapper[4836]: I0122 09:55:53.899585 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd" gracePeriod=600 Jan 22 09:55:54 crc kubenswrapper[4836]: I0122 09:55:54.151996 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd" exitCode=0 Jan 22 09:55:54 crc kubenswrapper[4836]: I0122 09:55:54.152086 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd"} Jan 22 09:55:54 crc kubenswrapper[4836]: I0122 09:55:54.152134 4836 scope.go:117] "RemoveContainer" containerID="117782c3d21842775fb6e77847fa130eb3331beedcb8f28d6e7af01ecb33a401" Jan 22 09:55:55 crc kubenswrapper[4836]: I0122 09:55:55.162922 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7"} Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.043431 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wblv2"] Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045502 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-controller" containerID="cri-o://21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045530 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="northd" containerID="cri-o://3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045717 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="sbdb" containerID="cri-o://93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045763 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="nbdb" containerID="cri-o://afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045676 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-acl-logging" containerID="cri-o://34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.045646 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.049506 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-node" containerID="cri-o://d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de" gracePeriod=30 Jan 22 09:56:52 crc kubenswrapper[4836]: I0122 09:56:52.082606 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" containerID="cri-o://ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba" gracePeriod=30 Jan 22 09:56:53 crc kubenswrapper[4836]: I0122 09:56:53.523976 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/3.log" Jan 22 09:56:53 crc kubenswrapper[4836]: I0122 09:56:53.526696 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-acl-logging/0.log" Jan 22 09:56:53 crc kubenswrapper[4836]: I0122 09:56:53.527465 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8" exitCode=143 Jan 22 09:56:53 crc kubenswrapper[4836]: I0122 09:56:53.527533 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.535551 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/2.log" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.536590 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/1.log" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.536663 4836 generic.go:334] "Generic (PLEG): container finished" podID="7fcb3e24-f305-4f39-af85-2dc05a0af79f" containerID="7ce729b6cbd16f82ef39e221e8774e4f11535f85358b4a0f4e713f3221ec39e9" exitCode=2 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.536746 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerDied","Data":"7ce729b6cbd16f82ef39e221e8774e4f11535f85358b4a0f4e713f3221ec39e9"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.536800 4836 scope.go:117] "RemoveContainer" containerID="023583bc9577f977e1a3e0a366e13462520c7e94ccff70ee914cb95bbb630633" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.537790 4836 scope.go:117] "RemoveContainer" containerID="7ce729b6cbd16f82ef39e221e8774e4f11535f85358b4a0f4e713f3221ec39e9" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.539992 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovnkube-controller/3.log" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.544382 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-acl-logging/0.log" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.545052 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-controller/0.log" Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546244 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba" exitCode=0 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546278 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4" exitCode=0 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546296 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c" exitCode=0 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546310 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f" exitCode=0 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546321 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de" exitCode=0 Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546320 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546381 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546424 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546450 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546468 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546482 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576"} Jan 22 09:56:54 crc kubenswrapper[4836]: I0122 09:56:54.546332 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576" exitCode=143 Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.019444 4836 scope.go:117] "RemoveContainer" containerID="f44a00bc7c2023cdeefa8b3c6b60f66b8e882afdbb687e2d6179b268a0e9fd2e" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.119970 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-acl-logging/0.log" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.120577 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-controller/0.log" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.120982 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175503 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175585 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175614 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175655 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175689 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175725 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175764 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175797 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175846 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175874 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175907 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175935 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.175978 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176014 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176062 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176092 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176152 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176191 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zbvn\" (UniqueName: \"kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176237 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176284 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch\") pod \"dc5486e7-b643-4b21-9be6-8d6f475b3908\" (UID: \"dc5486e7-b643-4b21-9be6-8d6f475b3908\") " Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176681 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176733 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176771 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log" (OuterVolumeSpecName: "node-log") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.176814 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket" (OuterVolumeSpecName: "log-socket") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177365 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177412 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177448 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177509 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177543 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.177956 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.178001 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179004 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash" (OuterVolumeSpecName: "host-slash") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179108 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179119 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179137 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179137 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.179919 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.182820 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gn46d"] Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.183807 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184030 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184081 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184171 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184188 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184206 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184218 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184230 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-node" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184242 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-node" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184262 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-acl-logging" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184274 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-acl-logging" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184291 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="northd" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184303 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="northd" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184321 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="sbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184331 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="sbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184468 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f60d64f9-1517-4093-9a99-abd333d12a36" containerName="registry" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184512 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f60d64f9-1517-4093-9a99-abd333d12a36" containerName="registry" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184531 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kubecfg-setup" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184544 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kubecfg-setup" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184559 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="nbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184570 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="nbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184586 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184597 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.184617 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184629 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184788 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184807 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="sbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184819 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f60d64f9-1517-4093-9a99-abd333d12a36" containerName="registry" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184835 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="nbdb" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184853 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-node" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184870 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184881 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184896 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184911 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovn-acl-logging" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184927 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184943 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184958 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="kube-rbac-proxy-ovn-metrics" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.184973 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="northd" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.185147 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.185164 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.185183 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.185193 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerName="ovnkube-controller" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.186379 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn" (OuterVolumeSpecName: "kube-api-access-4zbvn") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "kube-api-access-4zbvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.189629 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.199081 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "dc5486e7-b643-4b21-9be6-8d6f475b3908" (UID: "dc5486e7-b643-4b21-9be6-8d6f475b3908"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278103 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-slash\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278156 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-bin\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278180 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-var-lib-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278202 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-etc-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278223 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-netd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278308 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278348 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-kubelet\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278373 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovn-node-metrics-cert\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278400 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-script-lib\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278419 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-config\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278503 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-ovn\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278549 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-node-log\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278576 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghsp\" (UniqueName: \"kubernetes.io/projected/c7b8bb9c-dc71-4e83-9f36-8174dece720d-kube-api-access-rghsp\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278596 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-netns\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278622 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-log-socket\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278655 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-env-overrides\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278750 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-systemd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278857 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278917 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-systemd-units\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.278943 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279133 4836 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279155 4836 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279169 4836 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279182 4836 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279195 4836 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279208 4836 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279220 4836 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279232 4836 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279243 4836 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279276 4836 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279288 4836 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279303 4836 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-slash\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279330 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zbvn\" (UniqueName: \"kubernetes.io/projected/dc5486e7-b643-4b21-9be6-8d6f475b3908-kube-api-access-4zbvn\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279340 4836 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279348 4836 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279358 4836 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279366 4836 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-node-log\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279374 4836 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-log-socket\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279383 4836 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc5486e7-b643-4b21-9be6-8d6f475b3908-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.279390 4836 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc5486e7-b643-4b21-9be6-8d6f475b3908-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.379997 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-node-log\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380065 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghsp\" (UniqueName: \"kubernetes.io/projected/c7b8bb9c-dc71-4e83-9f36-8174dece720d-kube-api-access-rghsp\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380088 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-netns\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380106 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-log-socket\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380130 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-env-overrides\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380146 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-systemd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380157 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-node-log\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380192 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380196 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-log-socket\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380174 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-netns\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380168 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380221 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-systemd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380353 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-systemd-units\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380374 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380402 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-slash\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380417 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-bin\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380433 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-var-lib-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380439 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380437 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-systemd-units\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380453 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-etc-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380465 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-slash\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380474 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-var-lib-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380475 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-netd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380463 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-bin\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380502 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380509 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-cni-netd\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380500 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-etc-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380529 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-kubelet\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380553 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-host-kubelet\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380531 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-openvswitch\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380563 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovn-node-metrics-cert\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380633 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-script-lib\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380658 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-config\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380698 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-ovn\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.380775 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7b8bb9c-dc71-4e83-9f36-8174dece720d-run-ovn\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.381004 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-env-overrides\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.381368 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-script-lib\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.381572 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovnkube-config\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.383633 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7b8bb9c-dc71-4e83-9f36-8174dece720d-ovn-node-metrics-cert\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.395029 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghsp\" (UniqueName: \"kubernetes.io/projected/c7b8bb9c-dc71-4e83-9f36-8174dece720d-kube-api-access-rghsp\") pod \"ovnkube-node-gn46d\" (UID: \"c7b8bb9c-dc71-4e83-9f36-8174dece720d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.510645 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.554139 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/2.log" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.554293 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-89st4" event={"ID":"7fcb3e24-f305-4f39-af85-2dc05a0af79f","Type":"ContainerStarted","Data":"1a39358b0871552c532ad1a31ee15f20d32a3dc31e54d7c4a1e0395ccba7a9e3"} Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.558453 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-acl-logging/0.log" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.559361 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wblv2_dc5486e7-b643-4b21-9be6-8d6f475b3908/ovn-controller/0.log" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.560456 4836 generic.go:334] "Generic (PLEG): container finished" podID="dc5486e7-b643-4b21-9be6-8d6f475b3908" containerID="3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49" exitCode=0 Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.560530 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49"} Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.560554 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" event={"ID":"dc5486e7-b643-4b21-9be6-8d6f475b3908","Type":"ContainerDied","Data":"be5fbf6a39c5b074c978ee3d390727c7bc9d8187b20407412a81a3e8f12974a8"} Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.560571 4836 scope.go:117] "RemoveContainer" containerID="ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.560664 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wblv2" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.563240 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"d353d4b93640bd4f7227db85805ec106ea9b7f8cc55bd81426e63e40bd7d5d83"} Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.589565 4836 scope.go:117] "RemoveContainer" containerID="93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.605659 4836 scope.go:117] "RemoveContainer" containerID="afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.617839 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wblv2"] Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.622390 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wblv2"] Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.633279 4836 scope.go:117] "RemoveContainer" containerID="3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.645811 4836 scope.go:117] "RemoveContainer" containerID="b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.658305 4836 scope.go:117] "RemoveContainer" containerID="d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.726536 4836 scope.go:117] "RemoveContainer" containerID="34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.740388 4836 scope.go:117] "RemoveContainer" containerID="21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.754960 4836 scope.go:117] "RemoveContainer" containerID="db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.767318 4836 scope.go:117] "RemoveContainer" containerID="ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.767775 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba\": container with ID starting with ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba not found: ID does not exist" containerID="ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.767804 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba"} err="failed to get container status \"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba\": rpc error: code = NotFound desc = could not find container \"ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba\": container with ID starting with ffff576dc2efc49d8ff5235119671a1ad523e7fb1d3f27a2a6d7511d318a59ba not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.767825 4836 scope.go:117] "RemoveContainer" containerID="93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.768156 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\": container with ID starting with 93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4 not found: ID does not exist" containerID="93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768183 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4"} err="failed to get container status \"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\": rpc error: code = NotFound desc = could not find container \"93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4\": container with ID starting with 93e69aaed48694067f0fb9cb1f56a028d64e62554b81cff1de6821701a0199f4 not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768195 4836 scope.go:117] "RemoveContainer" containerID="afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.768473 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\": container with ID starting with afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c not found: ID does not exist" containerID="afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768531 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c"} err="failed to get container status \"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\": rpc error: code = NotFound desc = could not find container \"afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c\": container with ID starting with afd8af930a56ae74c32c2908f72cf7d3997923ff80f211c7594f3da0994d551c not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768572 4836 scope.go:117] "RemoveContainer" containerID="3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.768885 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\": container with ID starting with 3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49 not found: ID does not exist" containerID="3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768937 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49"} err="failed to get container status \"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\": rpc error: code = NotFound desc = could not find container \"3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49\": container with ID starting with 3daedf958663a4658a3438d8535feac61ef845c3e309b9fb90c44a0fa0e78d49 not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.768976 4836 scope.go:117] "RemoveContainer" containerID="b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.769355 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\": container with ID starting with b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f not found: ID does not exist" containerID="b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769381 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f"} err="failed to get container status \"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\": rpc error: code = NotFound desc = could not find container \"b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f\": container with ID starting with b73a2012bb72a4e3033b4e8566b4e4a9d31c0b4a8104f2151c0d15a09d74420f not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769396 4836 scope.go:117] "RemoveContainer" containerID="d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.769644 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\": container with ID starting with d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de not found: ID does not exist" containerID="d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769678 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de"} err="failed to get container status \"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\": rpc error: code = NotFound desc = could not find container \"d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de\": container with ID starting with d9bfee276ef23b4e83bde7a8c516267fece7ada3e24bf7f2b0473f9395b550de not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769701 4836 scope.go:117] "RemoveContainer" containerID="34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.769933 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\": container with ID starting with 34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8 not found: ID does not exist" containerID="34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769966 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8"} err="failed to get container status \"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\": rpc error: code = NotFound desc = could not find container \"34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8\": container with ID starting with 34de1b64ea1c44e5ccb909e22722e6848a26a36344a6124ed44111355e483bb8 not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.769989 4836 scope.go:117] "RemoveContainer" containerID="21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.770260 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\": container with ID starting with 21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576 not found: ID does not exist" containerID="21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.770288 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576"} err="failed to get container status \"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\": rpc error: code = NotFound desc = could not find container \"21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576\": container with ID starting with 21d4f89e5b33dfba727bb0d38b46cf98360e4bc8318ad2775a82dd1b925fa576 not found: ID does not exist" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.770312 4836 scope.go:117] "RemoveContainer" containerID="db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97" Jan 22 09:56:55 crc kubenswrapper[4836]: E0122 09:56:55.770671 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\": container with ID starting with db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97 not found: ID does not exist" containerID="db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97" Jan 22 09:56:55 crc kubenswrapper[4836]: I0122 09:56:55.770704 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97"} err="failed to get container status \"db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\": rpc error: code = NotFound desc = could not find container \"db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97\": container with ID starting with db1f2559eb962028180bb5ea307d6037ec0fb8fc7d24b17fbd12c31b7d78ea97 not found: ID does not exist" Jan 22 09:56:56 crc kubenswrapper[4836]: I0122 09:56:56.570104 4836 generic.go:334] "Generic (PLEG): container finished" podID="c7b8bb9c-dc71-4e83-9f36-8174dece720d" containerID="6588669d4532d97286781ac110ba46f39afe13bdeb58a3c16f018225e809eb02" exitCode=0 Jan 22 09:56:56 crc kubenswrapper[4836]: I0122 09:56:56.570171 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerDied","Data":"6588669d4532d97286781ac110ba46f39afe13bdeb58a3c16f018225e809eb02"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.004075 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5486e7-b643-4b21-9be6-8d6f475b3908" path="/var/lib/kubelet/pods/dc5486e7-b643-4b21-9be6-8d6f475b3908/volumes" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581639 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"c2c0f268497c868bf4dd6fee012b3b932d8ee7b506e5c3de16985f6e57ffbcb4"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581692 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"2fc70a1cdc310c5dcabee6a2814ab887bea4d91c2d43b064818305edc3d8d7df"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581705 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"cbb181f1482c323f78c8681c534d2477f79d160199140b22a7937186259d07ea"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581717 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"7c57da9365243e7ecbd9ae12b7b0ee82ba39d024584cd0e92e9f05b24ea3f7a7"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581728 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"80a884ece1bd7fc9e961b32f67e7c12babfab3f2ea6d3f072fafd9262e5c2ba8"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.581738 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"8beb338219530950b5301d7edb0c06c2a13c7beb75c0aed25dcf85db4234b6eb"} Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.840007 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-4d29j"] Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.840668 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.843010 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.843182 4836 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-z9ptr" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.843319 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.843329 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.919581 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.919660 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kgcc\" (UniqueName: \"kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:57 crc kubenswrapper[4836]: I0122 09:56:57.919770 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.020877 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.021395 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.021216 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.021437 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kgcc\" (UniqueName: \"kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.022246 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.038087 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kgcc\" (UniqueName: \"kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc\") pod \"crc-storage-crc-4d29j\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: I0122 09:56:58.158847 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: E0122 09:56:58.179819 4836 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(0c53c883a542be258be1932df7a3184bef6335ab4082846b701e1373ac7fc854): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 09:56:58 crc kubenswrapper[4836]: E0122 09:56:58.179896 4836 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(0c53c883a542be258be1932df7a3184bef6335ab4082846b701e1373ac7fc854): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: E0122 09:56:58.179919 4836 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(0c53c883a542be258be1932df7a3184bef6335ab4082846b701e1373ac7fc854): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:56:58 crc kubenswrapper[4836]: E0122 09:56:58.179973 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4d29j_crc-storage(7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4d29j_crc-storage(7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(0c53c883a542be258be1932df7a3184bef6335ab4082846b701e1373ac7fc854): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4d29j" podUID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" Jan 22 09:57:00 crc kubenswrapper[4836]: I0122 09:57:00.601281 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"5987e7549d85501077ae70e108642a2fa0e6f8c2e613978cdf951caa653898e2"} Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.622937 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" event={"ID":"c7b8bb9c-dc71-4e83-9f36-8174dece720d","Type":"ContainerStarted","Data":"55ff4f6d16ac91c2dece0df6852b85a233ff3d35972924b069153e8c0a24bfd5"} Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.623574 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.660028 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" podStartSLOduration=8.659998664 podStartE2EDuration="8.659998664s" podCreationTimestamp="2026-01-22 09:56:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:57:03.654991388 +0000 UTC m=+907.014759925" watchObservedRunningTime="2026-01-22 09:57:03.659998664 +0000 UTC m=+907.019767211" Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.663978 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.687351 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4d29j"] Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.687558 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:03 crc kubenswrapper[4836]: I0122 09:57:03.688226 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:03 crc kubenswrapper[4836]: E0122 09:57:03.711156 4836 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(5fd43ff70f1f0c71b5927975c4d450aedc1baf68110ad3d19a92cec565146f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 22 09:57:03 crc kubenswrapper[4836]: E0122 09:57:03.711246 4836 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(5fd43ff70f1f0c71b5927975c4d450aedc1baf68110ad3d19a92cec565146f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:03 crc kubenswrapper[4836]: E0122 09:57:03.711279 4836 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(5fd43ff70f1f0c71b5927975c4d450aedc1baf68110ad3d19a92cec565146f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:03 crc kubenswrapper[4836]: E0122 09:57:03.711345 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-4d29j_crc-storage(7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-4d29j_crc-storage(7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-4d29j_crc-storage_7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e_0(5fd43ff70f1f0c71b5927975c4d450aedc1baf68110ad3d19a92cec565146f3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-4d29j" podUID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" Jan 22 09:57:04 crc kubenswrapper[4836]: I0122 09:57:04.630709 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:04 crc kubenswrapper[4836]: I0122 09:57:04.631133 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:04 crc kubenswrapper[4836]: I0122 09:57:04.662800 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:16 crc kubenswrapper[4836]: I0122 09:57:16.997003 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:17 crc kubenswrapper[4836]: I0122 09:57:16.999986 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:17 crc kubenswrapper[4836]: I0122 09:57:17.156927 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4d29j"] Jan 22 09:57:17 crc kubenswrapper[4836]: I0122 09:57:17.167062 4836 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 09:57:17 crc kubenswrapper[4836]: I0122 09:57:17.702927 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4d29j" event={"ID":"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e","Type":"ContainerStarted","Data":"e86d516b06d2ce160bc2f5b86aa3e25c9ea5fa51a983bea5ea9332705bb54a25"} Jan 22 09:57:18 crc kubenswrapper[4836]: I0122 09:57:18.711367 4836 generic.go:334] "Generic (PLEG): container finished" podID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" containerID="070105d978182d2cb29abf24db3d3c5feb92aadd2a48fa6f79f98ce93307879f" exitCode=0 Jan 22 09:57:18 crc kubenswrapper[4836]: I0122 09:57:18.711479 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4d29j" event={"ID":"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e","Type":"ContainerDied","Data":"070105d978182d2cb29abf24db3d3c5feb92aadd2a48fa6f79f98ce93307879f"} Jan 22 09:57:19 crc kubenswrapper[4836]: I0122 09:57:19.943009 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.003626 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt\") pod \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.003730 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage\") pod \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.003773 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kgcc\" (UniqueName: \"kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc\") pod \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\" (UID: \"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e\") " Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.003785 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" (UID: "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.003956 4836 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.009360 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc" (OuterVolumeSpecName: "kube-api-access-5kgcc") pod "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" (UID: "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e"). InnerVolumeSpecName "kube-api-access-5kgcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.018206 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" (UID: "7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.105796 4836 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.105863 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kgcc\" (UniqueName: \"kubernetes.io/projected/7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e-kube-api-access-5kgcc\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.724276 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4d29j" event={"ID":"7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e","Type":"ContainerDied","Data":"e86d516b06d2ce160bc2f5b86aa3e25c9ea5fa51a983bea5ea9332705bb54a25"} Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.724319 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e86d516b06d2ce160bc2f5b86aa3e25c9ea5fa51a983bea5ea9332705bb54a25" Jan 22 09:57:20 crc kubenswrapper[4836]: I0122 09:57:20.724389 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4d29j" Jan 22 09:57:25 crc kubenswrapper[4836]: I0122 09:57:25.534615 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gn46d" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.823623 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2"] Jan 22 09:57:26 crc kubenswrapper[4836]: E0122 09:57:26.824111 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" containerName="storage" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.824125 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" containerName="storage" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.824219 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c8c0be8-7b5c-4779-b7fa-76b7bb0d772e" containerName="storage" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.825082 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.828262 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.835623 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2"] Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.884314 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.884541 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.884606 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7tdg\" (UniqueName: \"kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.986196 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.986257 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7tdg\" (UniqueName: \"kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.986291 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.986784 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:26 crc kubenswrapper[4836]: I0122 09:57:26.986834 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:27 crc kubenswrapper[4836]: I0122 09:57:27.005496 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7tdg\" (UniqueName: \"kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:27 crc kubenswrapper[4836]: I0122 09:57:27.144744 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:27 crc kubenswrapper[4836]: I0122 09:57:27.318190 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2"] Jan 22 09:57:27 crc kubenswrapper[4836]: I0122 09:57:27.763916 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerStarted","Data":"d53c2eaf347f00c960b5c038c17838dcaa94b9e957b60223ea228d552d1b7032"} Jan 22 09:57:27 crc kubenswrapper[4836]: I0122 09:57:27.763970 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerStarted","Data":"2395d854585e5886eef6928fbd794b94f1faa346c9890c595de2d35c64ec2abc"} Jan 22 09:57:28 crc kubenswrapper[4836]: I0122 09:57:28.771702 4836 generic.go:334] "Generic (PLEG): container finished" podID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerID="d53c2eaf347f00c960b5c038c17838dcaa94b9e957b60223ea228d552d1b7032" exitCode=0 Jan 22 09:57:28 crc kubenswrapper[4836]: I0122 09:57:28.771812 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerDied","Data":"d53c2eaf347f00c960b5c038c17838dcaa94b9e957b60223ea228d552d1b7032"} Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.164565 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.166241 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.186276 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.212118 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.212174 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd7kl\" (UniqueName: \"kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.212228 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.312811 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.312867 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.312894 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd7kl\" (UniqueName: \"kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.313309 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.313525 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.335733 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd7kl\" (UniqueName: \"kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl\") pod \"redhat-operators-c6pkc\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.491430 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.755549 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:29 crc kubenswrapper[4836]: W0122 09:57:29.763172 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfa0565a_e20a_43e5_8249_81bd45bb6ad7.slice/crio-a7d185095878214c73bb02ae3f4fe580aae064ed40e55bf93a40966f29d2ee38 WatchSource:0}: Error finding container a7d185095878214c73bb02ae3f4fe580aae064ed40e55bf93a40966f29d2ee38: Status 404 returned error can't find the container with id a7d185095878214c73bb02ae3f4fe580aae064ed40e55bf93a40966f29d2ee38 Jan 22 09:57:29 crc kubenswrapper[4836]: I0122 09:57:29.780811 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerStarted","Data":"a7d185095878214c73bb02ae3f4fe580aae064ed40e55bf93a40966f29d2ee38"} Jan 22 09:57:30 crc kubenswrapper[4836]: I0122 09:57:30.787474 4836 generic.go:334] "Generic (PLEG): container finished" podID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerID="33c8bdb7b905475b8a15a96f917381a8d99c1eeed1c1e0e009a268fbb5d123cf" exitCode=0 Jan 22 09:57:30 crc kubenswrapper[4836]: I0122 09:57:30.787562 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerDied","Data":"33c8bdb7b905475b8a15a96f917381a8d99c1eeed1c1e0e009a268fbb5d123cf"} Jan 22 09:57:30 crc kubenswrapper[4836]: I0122 09:57:30.791584 4836 generic.go:334] "Generic (PLEG): container finished" podID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerID="ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0" exitCode=0 Jan 22 09:57:30 crc kubenswrapper[4836]: I0122 09:57:30.791743 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerDied","Data":"ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0"} Jan 22 09:57:31 crc kubenswrapper[4836]: I0122 09:57:31.799244 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerStarted","Data":"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a"} Jan 22 09:57:31 crc kubenswrapper[4836]: I0122 09:57:31.809534 4836 generic.go:334] "Generic (PLEG): container finished" podID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerID="a89d5d6d48657e06c184b28e9e229aeef8975e4eaac7f71afb9be1b9a083bb20" exitCode=0 Jan 22 09:57:31 crc kubenswrapper[4836]: I0122 09:57:31.809731 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerDied","Data":"a89d5d6d48657e06c184b28e9e229aeef8975e4eaac7f71afb9be1b9a083bb20"} Jan 22 09:57:32 crc kubenswrapper[4836]: I0122 09:57:32.818973 4836 generic.go:334] "Generic (PLEG): container finished" podID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerID="87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a" exitCode=0 Jan 22 09:57:32 crc kubenswrapper[4836]: I0122 09:57:32.819070 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerDied","Data":"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a"} Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.029454 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.056761 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util\") pod \"37685439-5383-47b3-8cd0-ad982b0a7b5e\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.056858 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7tdg\" (UniqueName: \"kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg\") pod \"37685439-5383-47b3-8cd0-ad982b0a7b5e\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.056913 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle\") pod \"37685439-5383-47b3-8cd0-ad982b0a7b5e\" (UID: \"37685439-5383-47b3-8cd0-ad982b0a7b5e\") " Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.057569 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle" (OuterVolumeSpecName: "bundle") pod "37685439-5383-47b3-8cd0-ad982b0a7b5e" (UID: "37685439-5383-47b3-8cd0-ad982b0a7b5e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.059091 4836 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.061761 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg" (OuterVolumeSpecName: "kube-api-access-l7tdg") pod "37685439-5383-47b3-8cd0-ad982b0a7b5e" (UID: "37685439-5383-47b3-8cd0-ad982b0a7b5e"). InnerVolumeSpecName "kube-api-access-l7tdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.151946 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util" (OuterVolumeSpecName: "util") pod "37685439-5383-47b3-8cd0-ad982b0a7b5e" (UID: "37685439-5383-47b3-8cd0-ad982b0a7b5e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.160063 4836 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37685439-5383-47b3-8cd0-ad982b0a7b5e-util\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.160091 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7tdg\" (UniqueName: \"kubernetes.io/projected/37685439-5383-47b3-8cd0-ad982b0a7b5e-kube-api-access-l7tdg\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.830358 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" event={"ID":"37685439-5383-47b3-8cd0-ad982b0a7b5e","Type":"ContainerDied","Data":"2395d854585e5886eef6928fbd794b94f1faa346c9890c595de2d35c64ec2abc"} Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.831140 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2395d854585e5886eef6928fbd794b94f1faa346c9890c595de2d35c64ec2abc" Jan 22 09:57:33 crc kubenswrapper[4836]: I0122 09:57:33.830455 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2" Jan 22 09:57:34 crc kubenswrapper[4836]: I0122 09:57:34.840267 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerStarted","Data":"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f"} Jan 22 09:57:34 crc kubenswrapper[4836]: I0122 09:57:34.868470 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6pkc" podStartSLOduration=2.56763461 podStartE2EDuration="5.868440817s" podCreationTimestamp="2026-01-22 09:57:29 +0000 UTC" firstStartedPulling="2026-01-22 09:57:30.792640043 +0000 UTC m=+934.152408560" lastFinishedPulling="2026-01-22 09:57:34.09344625 +0000 UTC m=+937.453214767" observedRunningTime="2026-01-22 09:57:34.867288017 +0000 UTC m=+938.227056534" watchObservedRunningTime="2026-01-22 09:57:34.868440817 +0000 UTC m=+938.228209384" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.328514 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-nwcbn"] Jan 22 09:57:37 crc kubenswrapper[4836]: E0122 09:57:37.328758 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="util" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.328773 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="util" Jan 22 09:57:37 crc kubenswrapper[4836]: E0122 09:57:37.328789 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="extract" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.328797 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="extract" Jan 22 09:57:37 crc kubenswrapper[4836]: E0122 09:57:37.328808 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="pull" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.328815 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="pull" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.328927 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="37685439-5383-47b3-8cd0-ad982b0a7b5e" containerName="extract" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.329331 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.346883 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.347166 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.347441 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-cbwj4" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.362704 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-nwcbn"] Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.412752 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjxqm\" (UniqueName: \"kubernetes.io/projected/e6db6563-9866-4671-9f24-802e8717b5d7-kube-api-access-tjxqm\") pod \"nmstate-operator-646758c888-nwcbn\" (UID: \"e6db6563-9866-4671-9f24-802e8717b5d7\") " pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.513990 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjxqm\" (UniqueName: \"kubernetes.io/projected/e6db6563-9866-4671-9f24-802e8717b5d7-kube-api-access-tjxqm\") pod \"nmstate-operator-646758c888-nwcbn\" (UID: \"e6db6563-9866-4671-9f24-802e8717b5d7\") " pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.536474 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjxqm\" (UniqueName: \"kubernetes.io/projected/e6db6563-9866-4671-9f24-802e8717b5d7-kube-api-access-tjxqm\") pod \"nmstate-operator-646758c888-nwcbn\" (UID: \"e6db6563-9866-4671-9f24-802e8717b5d7\") " pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" Jan 22 09:57:37 crc kubenswrapper[4836]: I0122 09:57:37.662891 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" Jan 22 09:57:38 crc kubenswrapper[4836]: I0122 09:57:38.029502 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-nwcbn"] Jan 22 09:57:38 crc kubenswrapper[4836]: I0122 09:57:38.861633 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" event={"ID":"e6db6563-9866-4671-9f24-802e8717b5d7","Type":"ContainerStarted","Data":"98e3241176cb708fd6146c591e5df446435275ee03748b2ff9c9ca8cf14a8f74"} Jan 22 09:57:39 crc kubenswrapper[4836]: I0122 09:57:39.492022 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:39 crc kubenswrapper[4836]: I0122 09:57:39.492367 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:40 crc kubenswrapper[4836]: I0122 09:57:40.549814 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6pkc" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="registry-server" probeResult="failure" output=< Jan 22 09:57:40 crc kubenswrapper[4836]: timeout: failed to connect service ":50051" within 1s Jan 22 09:57:40 crc kubenswrapper[4836]: > Jan 22 09:57:49 crc kubenswrapper[4836]: I0122 09:57:49.582200 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:49 crc kubenswrapper[4836]: I0122 09:57:49.651016 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:49 crc kubenswrapper[4836]: I0122 09:57:49.816641 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:50 crc kubenswrapper[4836]: I0122 09:57:50.959250 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" event={"ID":"e6db6563-9866-4671-9f24-802e8717b5d7","Type":"ContainerStarted","Data":"d7be1b5e20de2f87ca8c46ae1701345f74fc8dd1449f00cce9c6b31daab6a1e9"} Jan 22 09:57:50 crc kubenswrapper[4836]: I0122 09:57:50.959669 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6pkc" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="registry-server" containerID="cri-o://7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f" gracePeriod=2 Jan 22 09:57:50 crc kubenswrapper[4836]: I0122 09:57:50.980076 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-nwcbn" podStartSLOduration=2.057611582 podStartE2EDuration="13.980059047s" podCreationTimestamp="2026-01-22 09:57:37 +0000 UTC" firstStartedPulling="2026-01-22 09:57:38.037742689 +0000 UTC m=+941.397511206" lastFinishedPulling="2026-01-22 09:57:49.960190154 +0000 UTC m=+953.319958671" observedRunningTime="2026-01-22 09:57:50.978871136 +0000 UTC m=+954.338639663" watchObservedRunningTime="2026-01-22 09:57:50.980059047 +0000 UTC m=+954.339827564" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.566924 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.619908 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities\") pod \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.620017 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content\") pod \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.620140 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd7kl\" (UniqueName: \"kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl\") pod \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\" (UID: \"cfa0565a-e20a-43e5-8249-81bd45bb6ad7\") " Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.620891 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities" (OuterVolumeSpecName: "utilities") pod "cfa0565a-e20a-43e5-8249-81bd45bb6ad7" (UID: "cfa0565a-e20a-43e5-8249-81bd45bb6ad7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.628297 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl" (OuterVolumeSpecName: "kube-api-access-zd7kl") pod "cfa0565a-e20a-43e5-8249-81bd45bb6ad7" (UID: "cfa0565a-e20a-43e5-8249-81bd45bb6ad7"). InnerVolumeSpecName "kube-api-access-zd7kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.722126 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.722162 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd7kl\" (UniqueName: \"kubernetes.io/projected/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-kube-api-access-zd7kl\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.749573 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfa0565a-e20a-43e5-8249-81bd45bb6ad7" (UID: "cfa0565a-e20a-43e5-8249-81bd45bb6ad7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.823129 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfa0565a-e20a-43e5-8249-81bd45bb6ad7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.883469 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-c8v9m"] Jan 22 09:57:51 crc kubenswrapper[4836]: E0122 09:57:51.883701 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="registry-server" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.883717 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="registry-server" Jan 22 09:57:51 crc kubenswrapper[4836]: E0122 09:57:51.883727 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="extract-content" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.883735 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="extract-content" Jan 22 09:57:51 crc kubenswrapper[4836]: E0122 09:57:51.883756 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="extract-utilities" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.883766 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="extract-utilities" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.883875 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerName="registry-server" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.884527 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.886874 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bfbpc" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.893715 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-c8v9m"] Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.897125 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j"] Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.897812 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.900115 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.918338 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-59bdc"] Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.919073 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.924051 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xqzv\" (UniqueName: \"kubernetes.io/projected/71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9-kube-api-access-7xqzv\") pod \"nmstate-metrics-54757c584b-c8v9m\" (UID: \"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.930234 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j"] Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.968410 4836 generic.go:334] "Generic (PLEG): container finished" podID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" containerID="7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f" exitCode=0 Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.968815 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6pkc" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.968934 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerDied","Data":"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f"} Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.969004 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6pkc" event={"ID":"cfa0565a-e20a-43e5-8249-81bd45bb6ad7","Type":"ContainerDied","Data":"a7d185095878214c73bb02ae3f4fe580aae064ed40e55bf93a40966f29d2ee38"} Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.969030 4836 scope.go:117] "RemoveContainer" containerID="7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f" Jan 22 09:57:51 crc kubenswrapper[4836]: I0122 09:57:51.995645 4836 scope.go:117] "RemoveContainer" containerID="87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025216 4836 scope.go:117] "RemoveContainer" containerID="ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025789 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-dbus-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025835 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xqzv\" (UniqueName: \"kubernetes.io/projected/71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9-kube-api-access-7xqzv\") pod \"nmstate-metrics-54757c584b-c8v9m\" (UID: \"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025859 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025886 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j6xn\" (UniqueName: \"kubernetes.io/projected/b89c6d12-f385-4e08-9a29-43befa6f48cc-kube-api-access-8j6xn\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025911 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-ovs-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025943 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5qz2\" (UniqueName: \"kubernetes.io/projected/63fcf02d-2d6a-4e47-8725-a4dde8fac447-kube-api-access-f5qz2\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.025960 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-nmstate-lock\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.047918 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.050881 4836 scope.go:117] "RemoveContainer" containerID="7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f" Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.052419 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f\": container with ID starting with 7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f not found: ID does not exist" containerID="7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.052475 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f"} err="failed to get container status \"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f\": rpc error: code = NotFound desc = could not find container \"7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f\": container with ID starting with 7efbdc0ac6addab498062b66cb0dec00b99609a163030ee72ab928dcdc4dec5f not found: ID does not exist" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.052509 4836 scope.go:117] "RemoveContainer" containerID="87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a" Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.052906 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a\": container with ID starting with 87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a not found: ID does not exist" containerID="87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.052953 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a"} err="failed to get container status \"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a\": rpc error: code = NotFound desc = could not find container \"87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a\": container with ID starting with 87d3c05849d03fae5b3e6573a142503063754a0390e4781c3e1ebb3319d0202a not found: ID does not exist" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.052987 4836 scope.go:117] "RemoveContainer" containerID="ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0" Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.053423 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0\": container with ID starting with ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0 not found: ID does not exist" containerID="ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.053458 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0"} err="failed to get container status \"ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0\": rpc error: code = NotFound desc = could not find container \"ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0\": container with ID starting with ff05faa09ad81c7f8cf7aadacf60a29124f04cdee652e6c26cf8d6df14d560f0 not found: ID does not exist" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.053944 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xqzv\" (UniqueName: \"kubernetes.io/projected/71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9-kube-api-access-7xqzv\") pod \"nmstate-metrics-54757c584b-c8v9m\" (UID: \"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.055392 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6pkc"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.083578 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.084173 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.085729 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.086345 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.090835 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-n7gxb" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.102313 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127209 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-ovs-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127325 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-ovs-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127377 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj2k7\" (UniqueName: \"kubernetes.io/projected/e97bdfd0-897c-45ac-9597-4d8c24e9d279-kube-api-access-qj2k7\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127408 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5qz2\" (UniqueName: \"kubernetes.io/projected/63fcf02d-2d6a-4e47-8725-a4dde8fac447-kube-api-access-f5qz2\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127621 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-nmstate-lock\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127426 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-nmstate-lock\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127784 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e97bdfd0-897c-45ac-9597-4d8c24e9d279-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127806 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-dbus-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127843 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127864 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.127894 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j6xn\" (UniqueName: \"kubernetes.io/projected/b89c6d12-f385-4e08-9a29-43befa6f48cc-kube-api-access-8j6xn\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.128294 4836 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.128343 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair podName:b89c6d12-f385-4e08-9a29-43befa6f48cc nodeName:}" failed. No retries permitted until 2026-01-22 09:57:52.628329055 +0000 UTC m=+955.988097572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-xw45j" (UID: "b89c6d12-f385-4e08-9a29-43befa6f48cc") : secret "openshift-nmstate-webhook" not found Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.128566 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/63fcf02d-2d6a-4e47-8725-a4dde8fac447-dbus-socket\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.144110 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j6xn\" (UniqueName: \"kubernetes.io/projected/b89c6d12-f385-4e08-9a29-43befa6f48cc-kube-api-access-8j6xn\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.145305 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5qz2\" (UniqueName: \"kubernetes.io/projected/63fcf02d-2d6a-4e47-8725-a4dde8fac447-kube-api-access-f5qz2\") pod \"nmstate-handler-59bdc\" (UID: \"63fcf02d-2d6a-4e47-8725-a4dde8fac447\") " pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.199302 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.229275 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e97bdfd0-897c-45ac-9597-4d8c24e9d279-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.229351 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.229420 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj2k7\" (UniqueName: \"kubernetes.io/projected/e97bdfd0-897c-45ac-9597-4d8c24e9d279-kube-api-access-qj2k7\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.229793 4836 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Jan 22 09:57:52 crc kubenswrapper[4836]: E0122 09:57:52.229853 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert podName:e97bdfd0-897c-45ac-9597-4d8c24e9d279 nodeName:}" failed. No retries permitted until 2026-01-22 09:57:52.729837375 +0000 UTC m=+956.089605892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-l29ws" (UID: "e97bdfd0-897c-45ac-9597-4d8c24e9d279") : secret "plugin-serving-cert" not found Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.230672 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e97bdfd0-897c-45ac-9597-4d8c24e9d279-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.237324 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.253032 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj2k7\" (UniqueName: \"kubernetes.io/projected/e97bdfd0-897c-45ac-9597-4d8c24e9d279-kube-api-access-qj2k7\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: W0122 09:57:52.278439 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63fcf02d_2d6a_4e47_8725_a4dde8fac447.slice/crio-a002d52408aae4df0e58c00389635762cca49f48cf285ce203a44b48f520095e WatchSource:0}: Error finding container a002d52408aae4df0e58c00389635762cca49f48cf285ce203a44b48f520095e: Status 404 returned error can't find the container with id a002d52408aae4df0e58c00389635762cca49f48cf285ce203a44b48f520095e Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.290608 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-58845f6759-nmwng"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.291271 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.310322 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58845f6759-nmwng"] Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330061 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-oauth-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330109 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-trusted-ca-bundle\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330158 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330187 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-service-ca\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330217 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-console-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330272 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-oauth-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.330328 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9zdb\" (UniqueName: \"kubernetes.io/projected/1854a9c0-6e4f-422d-94a8-57501680cb97-kube-api-access-k9zdb\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434691 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434748 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-service-ca\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434777 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-console-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434819 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-oauth-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434847 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9zdb\" (UniqueName: \"kubernetes.io/projected/1854a9c0-6e4f-422d-94a8-57501680cb97-kube-api-access-k9zdb\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434884 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-oauth-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.434901 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-trusted-ca-bundle\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.435818 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-console-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.436302 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-service-ca\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.436487 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-oauth-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.436721 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1854a9c0-6e4f-422d-94a8-57501680cb97-trusted-ca-bundle\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.438426 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-oauth-config\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.439835 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1854a9c0-6e4f-422d-94a8-57501680cb97-console-serving-cert\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.454376 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-c8v9m"] Jan 22 09:57:52 crc kubenswrapper[4836]: W0122 09:57:52.457494 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71b03f9f_75b7_45e3_ac01_fdcc7c5a3fc9.slice/crio-449a117f9f28c8aa03e7af25183fc0f669282fdc726344a0df85be468d18bbf2 WatchSource:0}: Error finding container 449a117f9f28c8aa03e7af25183fc0f669282fdc726344a0df85be468d18bbf2: Status 404 returned error can't find the container with id 449a117f9f28c8aa03e7af25183fc0f669282fdc726344a0df85be468d18bbf2 Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.458089 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9zdb\" (UniqueName: \"kubernetes.io/projected/1854a9c0-6e4f-422d-94a8-57501680cb97-kube-api-access-k9zdb\") pod \"console-58845f6759-nmwng\" (UID: \"1854a9c0-6e4f-422d-94a8-57501680cb97\") " pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.608198 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.637489 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.644431 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b89c6d12-f385-4e08-9a29-43befa6f48cc-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-xw45j\" (UID: \"b89c6d12-f385-4e08-9a29-43befa6f48cc\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.738767 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.744108 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e97bdfd0-897c-45ac-9597-4d8c24e9d279-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-l29ws\" (UID: \"e97bdfd0-897c-45ac-9597-4d8c24e9d279\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.810269 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.854096 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58845f6759-nmwng"] Jan 22 09:57:52 crc kubenswrapper[4836]: W0122 09:57:52.861519 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1854a9c0_6e4f_422d_94a8_57501680cb97.slice/crio-9face5d0ee782624be479cd61f54a6b5f09b303e4c8192852bb29c3c63681d5d WatchSource:0}: Error finding container 9face5d0ee782624be479cd61f54a6b5f09b303e4c8192852bb29c3c63681d5d: Status 404 returned error can't find the container with id 9face5d0ee782624be479cd61f54a6b5f09b303e4c8192852bb29c3c63681d5d Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.979284 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-59bdc" event={"ID":"63fcf02d-2d6a-4e47-8725-a4dde8fac447","Type":"ContainerStarted","Data":"a002d52408aae4df0e58c00389635762cca49f48cf285ce203a44b48f520095e"} Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.980533 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" event={"ID":"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9","Type":"ContainerStarted","Data":"449a117f9f28c8aa03e7af25183fc0f669282fdc726344a0df85be468d18bbf2"} Jan 22 09:57:52 crc kubenswrapper[4836]: I0122 09:57:52.986759 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58845f6759-nmwng" event={"ID":"1854a9c0-6e4f-422d-94a8-57501680cb97","Type":"ContainerStarted","Data":"9face5d0ee782624be479cd61f54a6b5f09b303e4c8192852bb29c3c63681d5d"} Jan 22 09:57:53 crc kubenswrapper[4836]: W0122 09:57:53.003034 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb89c6d12_f385_4e08_9a29_43befa6f48cc.slice/crio-dc1d98123522ff4003635085e3b7ab17104e2a9c7b717282e53df8894f075ecc WatchSource:0}: Error finding container dc1d98123522ff4003635085e3b7ab17104e2a9c7b717282e53df8894f075ecc: Status 404 returned error can't find the container with id dc1d98123522ff4003635085e3b7ab17104e2a9c7b717282e53df8894f075ecc Jan 22 09:57:53 crc kubenswrapper[4836]: I0122 09:57:53.004580 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa0565a-e20a-43e5-8249-81bd45bb6ad7" path="/var/lib/kubelet/pods/cfa0565a-e20a-43e5-8249-81bd45bb6ad7/volumes" Jan 22 09:57:53 crc kubenswrapper[4836]: I0122 09:57:53.005484 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j"] Jan 22 09:57:53 crc kubenswrapper[4836]: I0122 09:57:53.018758 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" Jan 22 09:57:53 crc kubenswrapper[4836]: I0122 09:57:53.392822 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws"] Jan 22 09:57:53 crc kubenswrapper[4836]: I0122 09:57:53.998990 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" event={"ID":"e97bdfd0-897c-45ac-9597-4d8c24e9d279","Type":"ContainerStarted","Data":"561ec0d3ea655df7a04a82cebd6fb1712a2d326d67b68096fa832af2ca0f97bf"} Jan 22 09:57:54 crc kubenswrapper[4836]: I0122 09:57:54.000204 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58845f6759-nmwng" event={"ID":"1854a9c0-6e4f-422d-94a8-57501680cb97","Type":"ContainerStarted","Data":"288785b8bfa6d9c1dcf38b3c57eb7d09f18b0520da1c7a1d22550b881a3d3b7d"} Jan 22 09:57:54 crc kubenswrapper[4836]: I0122 09:57:54.001554 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" event={"ID":"b89c6d12-f385-4e08-9a29-43befa6f48cc","Type":"ContainerStarted","Data":"dc1d98123522ff4003635085e3b7ab17104e2a9c7b717282e53df8894f075ecc"} Jan 22 09:57:54 crc kubenswrapper[4836]: I0122 09:57:54.017070 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-58845f6759-nmwng" podStartSLOduration=2.01702607 podStartE2EDuration="2.01702607s" podCreationTimestamp="2026-01-22 09:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:57:54.016234329 +0000 UTC m=+957.376002856" watchObservedRunningTime="2026-01-22 09:57:54.01702607 +0000 UTC m=+957.376794587" Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.008767 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-59bdc" event={"ID":"63fcf02d-2d6a-4e47-8725-a4dde8fac447","Type":"ContainerStarted","Data":"d69e5f06e5d160bcc50f282589b45251efad60d3f97e075be5f64226e0bb71b2"} Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.009133 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.011952 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" event={"ID":"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9","Type":"ContainerStarted","Data":"d6804b4601f263dedc4fc9196aa1d2b3bbb69d9ddc822c5501054012fa7f95f5"} Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.015463 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" event={"ID":"b89c6d12-f385-4e08-9a29-43befa6f48cc","Type":"ContainerStarted","Data":"92ffc8acea53d30ef956ce4341d64a412c0b38efb5c65106d103f07d77a8c345"} Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.015750 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.029785 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-59bdc" podStartSLOduration=1.681000192 podStartE2EDuration="4.029758322s" podCreationTimestamp="2026-01-22 09:57:51 +0000 UTC" firstStartedPulling="2026-01-22 09:57:52.286930552 +0000 UTC m=+955.646699069" lastFinishedPulling="2026-01-22 09:57:54.635688672 +0000 UTC m=+957.995457199" observedRunningTime="2026-01-22 09:57:55.02263963 +0000 UTC m=+958.382408147" watchObservedRunningTime="2026-01-22 09:57:55.029758322 +0000 UTC m=+958.389526859" Jan 22 09:57:55 crc kubenswrapper[4836]: I0122 09:57:55.040932 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" podStartSLOduration=2.409384844 podStartE2EDuration="4.040905822s" podCreationTimestamp="2026-01-22 09:57:51 +0000 UTC" firstStartedPulling="2026-01-22 09:57:53.005182962 +0000 UTC m=+956.364951479" lastFinishedPulling="2026-01-22 09:57:54.63670393 +0000 UTC m=+957.996472457" observedRunningTime="2026-01-22 09:57:55.037443169 +0000 UTC m=+958.397211686" watchObservedRunningTime="2026-01-22 09:57:55.040905822 +0000 UTC m=+958.400674339" Jan 22 09:57:56 crc kubenswrapper[4836]: I0122 09:57:56.022762 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" event={"ID":"e97bdfd0-897c-45ac-9597-4d8c24e9d279","Type":"ContainerStarted","Data":"8ea956f2af75fc63600f012b6e8b096ff9fe70b90af6f0f504b9ca18d8777d9e"} Jan 22 09:57:56 crc kubenswrapper[4836]: I0122 09:57:56.043346 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-l29ws" podStartSLOduration=1.938678743 podStartE2EDuration="4.043327007s" podCreationTimestamp="2026-01-22 09:57:52 +0000 UTC" firstStartedPulling="2026-01-22 09:57:53.404088423 +0000 UTC m=+956.763856940" lastFinishedPulling="2026-01-22 09:57:55.508736687 +0000 UTC m=+958.868505204" observedRunningTime="2026-01-22 09:57:56.039611136 +0000 UTC m=+959.399379653" watchObservedRunningTime="2026-01-22 09:57:56.043327007 +0000 UTC m=+959.403095524" Jan 22 09:57:57 crc kubenswrapper[4836]: I0122 09:57:57.035212 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" event={"ID":"71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9","Type":"ContainerStarted","Data":"ecbe7df70679fcfce0ae49c9918253a1fd33c0d26c8ea9e19263fc725bd982ff"} Jan 22 09:57:57 crc kubenswrapper[4836]: I0122 09:57:57.067643 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-c8v9m" podStartSLOduration=1.839882346 podStartE2EDuration="6.067590628s" podCreationTimestamp="2026-01-22 09:57:51 +0000 UTC" firstStartedPulling="2026-01-22 09:57:52.459469573 +0000 UTC m=+955.819238090" lastFinishedPulling="2026-01-22 09:57:56.687177845 +0000 UTC m=+960.046946372" observedRunningTime="2026-01-22 09:57:57.054288651 +0000 UTC m=+960.414057208" watchObservedRunningTime="2026-01-22 09:57:57.067590628 +0000 UTC m=+960.427359155" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.269336 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-59bdc" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.609841 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.609927 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.616732 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.919219 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.923634 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.929761 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.968293 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rft46\" (UniqueName: \"kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.968508 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:02 crc kubenswrapper[4836]: I0122 09:58:02.968588 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.069962 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rft46\" (UniqueName: \"kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.070114 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.070152 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.070952 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.070965 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.113222 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rft46\" (UniqueName: \"kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46\") pod \"redhat-marketplace-skn9d\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.123960 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-58845f6759-nmwng" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.173977 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.264884 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:03 crc kubenswrapper[4836]: I0122 09:58:03.467311 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:04 crc kubenswrapper[4836]: I0122 09:58:04.127000 4836 generic.go:334] "Generic (PLEG): container finished" podID="7d997a5b-193d-42af-8800-fbed65138fe5" containerID="119c1e24b1e3ed7a0cc2f306f55bdd9c3ee2c800926484b98cf23758f01b9043" exitCode=0 Jan 22 09:58:04 crc kubenswrapper[4836]: I0122 09:58:04.127200 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerDied","Data":"119c1e24b1e3ed7a0cc2f306f55bdd9c3ee2c800926484b98cf23758f01b9043"} Jan 22 09:58:04 crc kubenswrapper[4836]: I0122 09:58:04.127339 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerStarted","Data":"1697217e0632565aa9915a4a0258f205ad2b78b7ad88d4182c81c4c9601693d6"} Jan 22 09:58:07 crc kubenswrapper[4836]: I0122 09:58:07.150729 4836 generic.go:334] "Generic (PLEG): container finished" podID="7d997a5b-193d-42af-8800-fbed65138fe5" containerID="eb6d7586ef35b9a63fc8a594dd435e5e068d30aa1616dab6f223e6a353684d2d" exitCode=0 Jan 22 09:58:07 crc kubenswrapper[4836]: I0122 09:58:07.150846 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerDied","Data":"eb6d7586ef35b9a63fc8a594dd435e5e068d30aa1616dab6f223e6a353684d2d"} Jan 22 09:58:11 crc kubenswrapper[4836]: I0122 09:58:11.429201 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerStarted","Data":"00af1b67c927e77c173a9311d2d670ff4b5d153c0c16c8265907b328722a2c6c"} Jan 22 09:58:11 crc kubenswrapper[4836]: I0122 09:58:11.451651 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-skn9d" podStartSLOduration=4.817632472 podStartE2EDuration="9.451633654s" podCreationTimestamp="2026-01-22 09:58:02 +0000 UTC" firstStartedPulling="2026-01-22 09:58:04.128602006 +0000 UTC m=+967.488370523" lastFinishedPulling="2026-01-22 09:58:08.762603158 +0000 UTC m=+972.122371705" observedRunningTime="2026-01-22 09:58:11.447986556 +0000 UTC m=+974.807755073" watchObservedRunningTime="2026-01-22 09:58:11.451633654 +0000 UTC m=+974.811402171" Jan 22 09:58:12 crc kubenswrapper[4836]: I0122 09:58:12.816410 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-xw45j" Jan 22 09:58:13 crc kubenswrapper[4836]: I0122 09:58:13.265801 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:13 crc kubenswrapper[4836]: I0122 09:58:13.265879 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:13 crc kubenswrapper[4836]: I0122 09:58:13.352207 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:14 crc kubenswrapper[4836]: I0122 09:58:14.496620 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:14 crc kubenswrapper[4836]: I0122 09:58:14.548011 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:16 crc kubenswrapper[4836]: I0122 09:58:16.469221 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-skn9d" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="registry-server" containerID="cri-o://00af1b67c927e77c173a9311d2d670ff4b5d153c0c16c8265907b328722a2c6c" gracePeriod=2 Jan 22 09:58:19 crc kubenswrapper[4836]: I0122 09:58:19.490464 4836 generic.go:334] "Generic (PLEG): container finished" podID="7d997a5b-193d-42af-8800-fbed65138fe5" containerID="00af1b67c927e77c173a9311d2d670ff4b5d153c0c16c8265907b328722a2c6c" exitCode=0 Jan 22 09:58:19 crc kubenswrapper[4836]: I0122 09:58:19.490570 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerDied","Data":"00af1b67c927e77c173a9311d2d670ff4b5d153c0c16c8265907b328722a2c6c"} Jan 22 09:58:22 crc kubenswrapper[4836]: I0122 09:58:22.982234 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:22 crc kubenswrapper[4836]: I0122 09:58:22.988896 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:22 crc kubenswrapper[4836]: I0122 09:58:22.989639 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.067157 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.076916 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content\") pod \"7d997a5b-193d-42af-8800-fbed65138fe5\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.077127 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities\") pod \"7d997a5b-193d-42af-8800-fbed65138fe5\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.077176 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rft46\" (UniqueName: \"kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46\") pod \"7d997a5b-193d-42af-8800-fbed65138fe5\" (UID: \"7d997a5b-193d-42af-8800-fbed65138fe5\") " Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.077371 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.077416 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb5kv\" (UniqueName: \"kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.077440 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.078028 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities" (OuterVolumeSpecName: "utilities") pod "7d997a5b-193d-42af-8800-fbed65138fe5" (UID: "7d997a5b-193d-42af-8800-fbed65138fe5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.089555 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46" (OuterVolumeSpecName: "kube-api-access-rft46") pod "7d997a5b-193d-42af-8800-fbed65138fe5" (UID: "7d997a5b-193d-42af-8800-fbed65138fe5"). InnerVolumeSpecName "kube-api-access-rft46". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.098398 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d997a5b-193d-42af-8800-fbed65138fe5" (UID: "7d997a5b-193d-42af-8800-fbed65138fe5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178820 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178879 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb5kv\" (UniqueName: \"kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178902 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178951 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178962 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d997a5b-193d-42af-8800-fbed65138fe5-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.178972 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rft46\" (UniqueName: \"kubernetes.io/projected/7d997a5b-193d-42af-8800-fbed65138fe5-kube-api-access-rft46\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.179338 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.179407 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.194724 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb5kv\" (UniqueName: \"kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv\") pod \"community-operators-8djfb\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.315727 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.519536 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-skn9d" event={"ID":"7d997a5b-193d-42af-8800-fbed65138fe5","Type":"ContainerDied","Data":"1697217e0632565aa9915a4a0258f205ad2b78b7ad88d4182c81c4c9601693d6"} Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.519584 4836 scope.go:117] "RemoveContainer" containerID="00af1b67c927e77c173a9311d2d670ff4b5d153c0c16c8265907b328722a2c6c" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.519696 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-skn9d" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.543514 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.553089 4836 scope.go:117] "RemoveContainer" containerID="eb6d7586ef35b9a63fc8a594dd435e5e068d30aa1616dab6f223e6a353684d2d" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.570339 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.575100 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-skn9d"] Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.596593 4836 scope.go:117] "RemoveContainer" containerID="119c1e24b1e3ed7a0cc2f306f55bdd9c3ee2c800926484b98cf23758f01b9043" Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.897847 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:58:23 crc kubenswrapper[4836]: I0122 09:58:23.898274 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:58:24 crc kubenswrapper[4836]: I0122 09:58:24.526671 4836 generic.go:334] "Generic (PLEG): container finished" podID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerID="9771606fb24fa9562edac953be76d9905167cd9c89b42aa75aa430dbfac6ad03" exitCode=0 Jan 22 09:58:24 crc kubenswrapper[4836]: I0122 09:58:24.526732 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerDied","Data":"9771606fb24fa9562edac953be76d9905167cd9c89b42aa75aa430dbfac6ad03"} Jan 22 09:58:24 crc kubenswrapper[4836]: I0122 09:58:24.527147 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerStarted","Data":"7381c39536db89d4068e7e908115fc0e23a183bc5e88320dcbee69ed1423160e"} Jan 22 09:58:25 crc kubenswrapper[4836]: I0122 09:58:25.006205 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" path="/var/lib/kubelet/pods/7d997a5b-193d-42af-8800-fbed65138fe5/volumes" Jan 22 09:58:27 crc kubenswrapper[4836]: I0122 09:58:27.562941 4836 generic.go:334] "Generic (PLEG): container finished" podID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerID="aa66abc98012d10147d632f909090ea9001a5877f56f81e805e36d4b44cd6980" exitCode=0 Jan 22 09:58:27 crc kubenswrapper[4836]: I0122 09:58:27.563085 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerDied","Data":"aa66abc98012d10147d632f909090ea9001a5877f56f81e805e36d4b44cd6980"} Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.222469 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-tbm5x" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" containerID="cri-o://4a5b7cccb4349646461681a49b4fef0afaad02d65742313b74187899f0f63635" gracePeriod=15 Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.580328 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-tbm5x_37f85569-398f-4d97-8971-958d0010e4d4/console/0.log" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.580628 4836 generic.go:334] "Generic (PLEG): container finished" podID="37f85569-398f-4d97-8971-958d0010e4d4" containerID="4a5b7cccb4349646461681a49b4fef0afaad02d65742313b74187899f0f63635" exitCode=2 Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.580657 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tbm5x" event={"ID":"37f85569-398f-4d97-8971-958d0010e4d4","Type":"ContainerDied","Data":"4a5b7cccb4349646461681a49b4fef0afaad02d65742313b74187899f0f63635"} Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.720290 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-tbm5x_37f85569-398f-4d97-8971-958d0010e4d4/console/0.log" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.720742 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.777231 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.777314 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.777361 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.777386 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5vq2\" (UniqueName: \"kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.778427 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.778456 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.778493 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config" (OuterVolumeSpecName: "console-config") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.778893 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.778512 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca\") pod \"37f85569-398f-4d97-8971-958d0010e4d4\" (UID: \"37f85569-398f-4d97-8971-958d0010e4d4\") " Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779401 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779435 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca" (OuterVolumeSpecName: "service-ca") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779676 4836 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-console-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779699 4836 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779708 4836 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.779718 4836 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/37f85569-398f-4d97-8971-958d0010e4d4-service-ca\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.783812 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.784716 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2" (OuterVolumeSpecName: "kube-api-access-c5vq2") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "kube-api-access-c5vq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.792338 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "37f85569-398f-4d97-8971-958d0010e4d4" (UID: "37f85569-398f-4d97-8971-958d0010e4d4"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.880658 4836 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.880702 4836 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/37f85569-398f-4d97-8971-958d0010e4d4-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:28 crc kubenswrapper[4836]: I0122 09:58:28.880713 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5vq2\" (UniqueName: \"kubernetes.io/projected/37f85569-398f-4d97-8971-958d0010e4d4-kube-api-access-c5vq2\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:29 crc kubenswrapper[4836]: E0122 09:58:29.042108 4836 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37f85569_398f_4d97_8971_958d0010e4d4.slice/crio-2b019d42321cb4461643884fa37721896518a0ae7fce4a6c4b27d630488e30d9\": RecentStats: unable to find data in memory cache]" Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.587823 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-tbm5x_37f85569-398f-4d97-8971-958d0010e4d4/console/0.log" Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.588158 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-tbm5x" event={"ID":"37f85569-398f-4d97-8971-958d0010e4d4","Type":"ContainerDied","Data":"2b019d42321cb4461643884fa37721896518a0ae7fce4a6c4b27d630488e30d9"} Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.588201 4836 scope.go:117] "RemoveContainer" containerID="4a5b7cccb4349646461681a49b4fef0afaad02d65742313b74187899f0f63635" Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.588316 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-tbm5x" Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.612124 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:58:29 crc kubenswrapper[4836]: I0122 09:58:29.616961 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-tbm5x"] Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.597619 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerStarted","Data":"74f980699b216343957569b48eeb1bb412c5ee598560cf577a5a5a247a85eb33"} Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731213 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk"] Jan 22 09:58:30 crc kubenswrapper[4836]: E0122 09:58:30.731553 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="extract-utilities" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731585 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="extract-utilities" Jan 22 09:58:30 crc kubenswrapper[4836]: E0122 09:58:30.731614 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731629 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" Jan 22 09:58:30 crc kubenswrapper[4836]: E0122 09:58:30.731652 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="extract-content" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731666 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="extract-content" Jan 22 09:58:30 crc kubenswrapper[4836]: E0122 09:58:30.731683 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="registry-server" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731699 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="registry-server" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731881 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="37f85569-398f-4d97-8971-958d0010e4d4" containerName="console" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.731927 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d997a5b-193d-42af-8800-fbed65138fe5" containerName="registry-server" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.733294 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.739894 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk"] Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.741178 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.808261 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.808329 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.808363 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s9k4\" (UniqueName: \"kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.909931 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.910163 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.910259 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s9k4\" (UniqueName: \"kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.910779 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.911355 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:30 crc kubenswrapper[4836]: I0122 09:58:30.931978 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s9k4\" (UniqueName: \"kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:31 crc kubenswrapper[4836]: I0122 09:58:31.009736 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37f85569-398f-4d97-8971-958d0010e4d4" path="/var/lib/kubelet/pods/37f85569-398f-4d97-8971-958d0010e4d4/volumes" Jan 22 09:58:31 crc kubenswrapper[4836]: I0122 09:58:31.050780 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:31 crc kubenswrapper[4836]: I0122 09:58:31.326414 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk"] Jan 22 09:58:31 crc kubenswrapper[4836]: I0122 09:58:31.607094 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerStarted","Data":"cf00da902fdfbcab2e9d92e4b7f6bd3ccb2f3c0914bdd1ebacddc30341fa434a"} Jan 22 09:58:31 crc kubenswrapper[4836]: I0122 09:58:31.630132 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8djfb" podStartSLOduration=4.587881618 podStartE2EDuration="9.63011566s" podCreationTimestamp="2026-01-22 09:58:22 +0000 UTC" firstStartedPulling="2026-01-22 09:58:24.529523936 +0000 UTC m=+987.889292453" lastFinishedPulling="2026-01-22 09:58:29.571757978 +0000 UTC m=+992.931526495" observedRunningTime="2026-01-22 09:58:31.62638994 +0000 UTC m=+994.986158477" watchObservedRunningTime="2026-01-22 09:58:31.63011566 +0000 UTC m=+994.989884177" Jan 22 09:58:32 crc kubenswrapper[4836]: I0122 09:58:32.617096 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerStarted","Data":"634c661da7afd92a35880bb5b310b55597bb94b69201accb9ffc24a74f774769"} Jan 22 09:58:33 crc kubenswrapper[4836]: I0122 09:58:33.316605 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:33 crc kubenswrapper[4836]: I0122 09:58:33.316933 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:33 crc kubenswrapper[4836]: I0122 09:58:33.378858 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:33 crc kubenswrapper[4836]: I0122 09:58:33.624518 4836 generic.go:334] "Generic (PLEG): container finished" podID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerID="634c661da7afd92a35880bb5b310b55597bb94b69201accb9ffc24a74f774769" exitCode=0 Jan 22 09:58:33 crc kubenswrapper[4836]: I0122 09:58:33.624566 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerDied","Data":"634c661da7afd92a35880bb5b310b55597bb94b69201accb9ffc24a74f774769"} Jan 22 09:58:35 crc kubenswrapper[4836]: I0122 09:58:35.637988 4836 generic.go:334] "Generic (PLEG): container finished" podID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerID="fd58ba0ef114481ab44879146e3ad53f35a5ba4ac4155fe1b155c141a1f7a21f" exitCode=0 Jan 22 09:58:35 crc kubenswrapper[4836]: I0122 09:58:35.638169 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerDied","Data":"fd58ba0ef114481ab44879146e3ad53f35a5ba4ac4155fe1b155c141a1f7a21f"} Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.645701 4836 generic.go:334] "Generic (PLEG): container finished" podID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerID="9e6267b4ed2782281b257bc04042a24654ccc3bb6bf92ff43f69f92624c18fc1" exitCode=0 Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.645779 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerDied","Data":"9e6267b4ed2782281b257bc04042a24654ccc3bb6bf92ff43f69f92624c18fc1"} Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.694408 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9bntn"] Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.696512 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.714458 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bntn"] Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.888898 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zflwh\" (UniqueName: \"kubernetes.io/projected/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-kube-api-access-zflwh\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.888987 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-utilities\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.889028 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-catalog-content\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.989979 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zflwh\" (UniqueName: \"kubernetes.io/projected/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-kube-api-access-zflwh\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.990355 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-utilities\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.990404 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-catalog-content\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.990920 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-utilities\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:36 crc kubenswrapper[4836]: I0122 09:58:36.991145 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-catalog-content\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:37 crc kubenswrapper[4836]: I0122 09:58:37.010399 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zflwh\" (UniqueName: \"kubernetes.io/projected/dfab7eb6-8f9f-44cd-8cc4-2e364698c767-kube-api-access-zflwh\") pod \"certified-operators-9bntn\" (UID: \"dfab7eb6-8f9f-44cd-8cc4-2e364698c767\") " pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:37 crc kubenswrapper[4836]: I0122 09:58:37.024457 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:58:37 crc kubenswrapper[4836]: I0122 09:58:37.531778 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bntn"] Jan 22 09:58:37 crc kubenswrapper[4836]: W0122 09:58:37.541710 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfab7eb6_8f9f_44cd_8cc4_2e364698c767.slice/crio-48513b0ecd9288c0cc1dcc9cb7b69b6db88e2b1456183c8b3cce0717f867a870 WatchSource:0}: Error finding container 48513b0ecd9288c0cc1dcc9cb7b69b6db88e2b1456183c8b3cce0717f867a870: Status 404 returned error can't find the container with id 48513b0ecd9288c0cc1dcc9cb7b69b6db88e2b1456183c8b3cce0717f867a870 Jan 22 09:58:37 crc kubenswrapper[4836]: I0122 09:58:37.650735 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bntn" event={"ID":"dfab7eb6-8f9f-44cd-8cc4-2e364698c767","Type":"ContainerStarted","Data":"48513b0ecd9288c0cc1dcc9cb7b69b6db88e2b1456183c8b3cce0717f867a870"} Jan 22 09:58:37 crc kubenswrapper[4836]: I0122 09:58:37.875594 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.004870 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util\") pod \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.005034 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s9k4\" (UniqueName: \"kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4\") pod \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.005166 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle\") pod \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\" (UID: \"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd\") " Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.006448 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle" (OuterVolumeSpecName: "bundle") pod "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" (UID: "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.013861 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4" (OuterVolumeSpecName: "kube-api-access-7s9k4") pod "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" (UID: "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd"). InnerVolumeSpecName "kube-api-access-7s9k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.106741 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s9k4\" (UniqueName: \"kubernetes.io/projected/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-kube-api-access-7s9k4\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.106799 4836 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.657274 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" event={"ID":"3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd","Type":"ContainerDied","Data":"cf00da902fdfbcab2e9d92e4b7f6bd3ccb2f3c0914bdd1ebacddc30341fa434a"} Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.657572 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf00da902fdfbcab2e9d92e4b7f6bd3ccb2f3c0914bdd1ebacddc30341fa434a" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.657414 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.830155 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util" (OuterVolumeSpecName: "util") pod "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" (UID: "3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:38 crc kubenswrapper[4836]: I0122 09:58:38.917890 4836 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd-util\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:41 crc kubenswrapper[4836]: I0122 09:58:41.677594 4836 generic.go:334] "Generic (PLEG): container finished" podID="dfab7eb6-8f9f-44cd-8cc4-2e364698c767" containerID="22b7f78427deef2b498f1a26dbade6d034e2d8ee909641624760623339bf8cea" exitCode=0 Jan 22 09:58:41 crc kubenswrapper[4836]: I0122 09:58:41.677644 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bntn" event={"ID":"dfab7eb6-8f9f-44cd-8cc4-2e364698c767","Type":"ContainerDied","Data":"22b7f78427deef2b498f1a26dbade6d034e2d8ee909641624760623339bf8cea"} Jan 22 09:58:43 crc kubenswrapper[4836]: I0122 09:58:43.383107 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:45 crc kubenswrapper[4836]: I0122 09:58:45.870674 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:45 crc kubenswrapper[4836]: I0122 09:58:45.870898 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8djfb" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="registry-server" containerID="cri-o://74f980699b216343957569b48eeb1bb412c5ee598560cf577a5a5a247a85eb33" gracePeriod=2 Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.238658 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp"] Jan 22 09:58:46 crc kubenswrapper[4836]: E0122 09:58:46.239142 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="pull" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.239153 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="pull" Jan 22 09:58:46 crc kubenswrapper[4836]: E0122 09:58:46.239166 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="util" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.239173 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="util" Jan 22 09:58:46 crc kubenswrapper[4836]: E0122 09:58:46.239179 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="extract" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.239185 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="extract" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.239278 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd" containerName="extract" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.239653 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.241106 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.241767 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.241870 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.246143 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.246352 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-298pb" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.257165 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp"] Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.342381 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nzm2\" (UniqueName: \"kubernetes.io/projected/acf969fa-26d3-4c92-9b0e-553cedfd3a04-kube-api-access-6nzm2\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.342438 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-webhook-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.342462 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-apiservice-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.443628 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nzm2\" (UniqueName: \"kubernetes.io/projected/acf969fa-26d3-4c92-9b0e-553cedfd3a04-kube-api-access-6nzm2\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.443691 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-webhook-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.443713 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-apiservice-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.449594 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-apiservice-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.464316 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/acf969fa-26d3-4c92-9b0e-553cedfd3a04-webhook-cert\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.465400 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nzm2\" (UniqueName: \"kubernetes.io/projected/acf969fa-26d3-4c92-9b0e-553cedfd3a04-kube-api-access-6nzm2\") pod \"metallb-operator-controller-manager-799d7d6dcc-bvjvp\" (UID: \"acf969fa-26d3-4c92-9b0e-553cedfd3a04\") " pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.488504 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn"] Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.489471 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.493451 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.493568 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-k7mdw" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.493700 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.516177 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn"] Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.558899 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.559332 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t29zp\" (UniqueName: \"kubernetes.io/projected/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-kube-api-access-t29zp\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.559382 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-webhook-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.559456 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-apiservice-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.660702 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t29zp\" (UniqueName: \"kubernetes.io/projected/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-kube-api-access-t29zp\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.660751 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-webhook-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.660821 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-apiservice-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.665642 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-apiservice-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.678447 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-webhook-cert\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.681420 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t29zp\" (UniqueName: \"kubernetes.io/projected/e785c5eb-f7d6-4c65-8632-16a4d7e4f161-kube-api-access-t29zp\") pod \"metallb-operator-webhook-server-7fd664ffb8-qplwn\" (UID: \"e785c5eb-f7d6-4c65-8632-16a4d7e4f161\") " pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.720964 4836 generic.go:334] "Generic (PLEG): container finished" podID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerID="74f980699b216343957569b48eeb1bb412c5ee598560cf577a5a5a247a85eb33" exitCode=0 Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.721007 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerDied","Data":"74f980699b216343957569b48eeb1bb412c5ee598560cf577a5a5a247a85eb33"} Jan 22 09:58:46 crc kubenswrapper[4836]: I0122 09:58:46.809351 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.007462 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.122271 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content\") pod \"91ace0d4-386b-486d-bef4-2a68a812d7a9\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.122419 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities\") pod \"91ace0d4-386b-486d-bef4-2a68a812d7a9\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.122461 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb5kv\" (UniqueName: \"kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv\") pod \"91ace0d4-386b-486d-bef4-2a68a812d7a9\" (UID: \"91ace0d4-386b-486d-bef4-2a68a812d7a9\") " Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.123136 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities" (OuterVolumeSpecName: "utilities") pod "91ace0d4-386b-486d-bef4-2a68a812d7a9" (UID: "91ace0d4-386b-486d-bef4-2a68a812d7a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.123476 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.144194 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv" (OuterVolumeSpecName: "kube-api-access-rb5kv") pod "91ace0d4-386b-486d-bef4-2a68a812d7a9" (UID: "91ace0d4-386b-486d-bef4-2a68a812d7a9"). InnerVolumeSpecName "kube-api-access-rb5kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.194902 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91ace0d4-386b-486d-bef4-2a68a812d7a9" (UID: "91ace0d4-386b-486d-bef4-2a68a812d7a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.226199 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb5kv\" (UniqueName: \"kubernetes.io/projected/91ace0d4-386b-486d-bef4-2a68a812d7a9-kube-api-access-rb5kv\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.226236 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ace0d4-386b-486d-bef4-2a68a812d7a9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.367452 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn"] Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.468757 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp"] Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.750485 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8djfb" event={"ID":"91ace0d4-386b-486d-bef4-2a68a812d7a9","Type":"ContainerDied","Data":"7381c39536db89d4068e7e908115fc0e23a183bc5e88320dcbee69ed1423160e"} Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.750554 4836 scope.go:117] "RemoveContainer" containerID="74f980699b216343957569b48eeb1bb412c5ee598560cf577a5a5a247a85eb33" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.750571 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8djfb" Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.751835 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" event={"ID":"e785c5eb-f7d6-4c65-8632-16a4d7e4f161","Type":"ContainerStarted","Data":"381072b47dcafc9c8cecb97a79da74965885f75fbed1f36555fb313d45f38346"} Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.795352 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:51 crc kubenswrapper[4836]: I0122 09:58:51.804307 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8djfb"] Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.003616 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" path="/var/lib/kubelet/pods/91ace0d4-386b-486d-bef4-2a68a812d7a9/volumes" Jan 22 09:58:53 crc kubenswrapper[4836]: W0122 09:58:53.518851 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacf969fa_26d3_4c92_9b0e_553cedfd3a04.slice/crio-3209733c0f42395fe2d3cc5024785ee73032448b2ab363c503a476f88640a092 WatchSource:0}: Error finding container 3209733c0f42395fe2d3cc5024785ee73032448b2ab363c503a476f88640a092: Status 404 returned error can't find the container with id 3209733c0f42395fe2d3cc5024785ee73032448b2ab363c503a476f88640a092 Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.536674 4836 scope.go:117] "RemoveContainer" containerID="aa66abc98012d10147d632f909090ea9001a5877f56f81e805e36d4b44cd6980" Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.553972 4836 scope.go:117] "RemoveContainer" containerID="9771606fb24fa9562edac953be76d9905167cd9c89b42aa75aa430dbfac6ad03" Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.765217 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" event={"ID":"acf969fa-26d3-4c92-9b0e-553cedfd3a04","Type":"ContainerStarted","Data":"3209733c0f42395fe2d3cc5024785ee73032448b2ab363c503a476f88640a092"} Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.898148 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:58:53 crc kubenswrapper[4836]: I0122 09:58:53.898220 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:58:54 crc kubenswrapper[4836]: I0122 09:58:54.772346 4836 generic.go:334] "Generic (PLEG): container finished" podID="dfab7eb6-8f9f-44cd-8cc4-2e364698c767" containerID="f8b55ccb8d1295c0c4065b4a8016ca0bfb56fbc90ef72d732e370b84ffa6bf4e" exitCode=0 Jan 22 09:58:54 crc kubenswrapper[4836]: I0122 09:58:54.772386 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bntn" event={"ID":"dfab7eb6-8f9f-44cd-8cc4-2e364698c767","Type":"ContainerDied","Data":"f8b55ccb8d1295c0c4065b4a8016ca0bfb56fbc90ef72d732e370b84ffa6bf4e"} Jan 22 09:58:58 crc kubenswrapper[4836]: I0122 09:58:58.798834 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9bntn" event={"ID":"dfab7eb6-8f9f-44cd-8cc4-2e364698c767","Type":"ContainerStarted","Data":"1558c843803b9b6727a584597f878ecb90c2da5105d8a31fd1ebc6e7cbc99b89"} Jan 22 09:58:58 crc kubenswrapper[4836]: I0122 09:58:58.820661 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9bntn" podStartSLOduration=9.252015808 podStartE2EDuration="22.820637728s" podCreationTimestamp="2026-01-22 09:58:36 +0000 UTC" firstStartedPulling="2026-01-22 09:58:41.678938311 +0000 UTC m=+1005.038706828" lastFinishedPulling="2026-01-22 09:58:55.247560241 +0000 UTC m=+1018.607328748" observedRunningTime="2026-01-22 09:58:58.818594663 +0000 UTC m=+1022.178363200" watchObservedRunningTime="2026-01-22 09:58:58.820637728 +0000 UTC m=+1022.180406245" Jan 22 09:59:04 crc kubenswrapper[4836]: I0122 09:59:04.849553 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" event={"ID":"acf969fa-26d3-4c92-9b0e-553cedfd3a04","Type":"ContainerStarted","Data":"0cfca634b6a26e80f6f6e522235e79f95d25446ac72d906a60bf10479c5af1e7"} Jan 22 09:59:04 crc kubenswrapper[4836]: I0122 09:59:04.859421 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" event={"ID":"e785c5eb-f7d6-4c65-8632-16a4d7e4f161","Type":"ContainerStarted","Data":"192d86eac4ec4c9b74674ed86b71bd95c15eb75d3ae343befde16b967184fed9"} Jan 22 09:59:05 crc kubenswrapper[4836]: I0122 09:59:05.865212 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:59:05 crc kubenswrapper[4836]: I0122 09:59:05.890390 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" podStartSLOduration=6.913021742 podStartE2EDuration="19.890367138s" podCreationTimestamp="2026-01-22 09:58:46 +0000 UTC" firstStartedPulling="2026-01-22 09:58:51.378386698 +0000 UTC m=+1014.738155255" lastFinishedPulling="2026-01-22 09:59:04.355732134 +0000 UTC m=+1027.715500651" observedRunningTime="2026-01-22 09:59:05.885423984 +0000 UTC m=+1029.245192501" watchObservedRunningTime="2026-01-22 09:59:05.890367138 +0000 UTC m=+1029.250135655" Jan 22 09:59:05 crc kubenswrapper[4836]: I0122 09:59:05.912108 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" podStartSLOduration=9.083621859 podStartE2EDuration="19.912088835s" podCreationTimestamp="2026-01-22 09:58:46 +0000 UTC" firstStartedPulling="2026-01-22 09:58:53.536704002 +0000 UTC m=+1016.896472519" lastFinishedPulling="2026-01-22 09:59:04.365170978 +0000 UTC m=+1027.724939495" observedRunningTime="2026-01-22 09:59:05.906400681 +0000 UTC m=+1029.266169238" watchObservedRunningTime="2026-01-22 09:59:05.912088835 +0000 UTC m=+1029.271857362" Jan 22 09:59:06 crc kubenswrapper[4836]: I0122 09:59:06.560218 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:59:07 crc kubenswrapper[4836]: I0122 09:59:07.025166 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:59:07 crc kubenswrapper[4836]: I0122 09:59:07.025246 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:59:07 crc kubenswrapper[4836]: I0122 09:59:07.065806 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:59:07 crc kubenswrapper[4836]: I0122 09:59:07.911776 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9bntn" Jan 22 09:59:07 crc kubenswrapper[4836]: I0122 09:59:07.969476 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9bntn"] Jan 22 09:59:08 crc kubenswrapper[4836]: I0122 09:59:08.006507 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:59:08 crc kubenswrapper[4836]: I0122 09:59:08.006807 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wxfgn" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="registry-server" containerID="cri-o://c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" gracePeriod=2 Jan 22 09:59:08 crc kubenswrapper[4836]: E0122 09:59:08.109776 4836 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91 is running failed: container process not found" containerID="c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" cmd=["grpc_health_probe","-addr=:50051"] Jan 22 09:59:08 crc kubenswrapper[4836]: E0122 09:59:08.110343 4836 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91 is running failed: container process not found" containerID="c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" cmd=["grpc_health_probe","-addr=:50051"] Jan 22 09:59:08 crc kubenswrapper[4836]: E0122 09:59:08.111005 4836 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91 is running failed: container process not found" containerID="c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" cmd=["grpc_health_probe","-addr=:50051"] Jan 22 09:59:08 crc kubenswrapper[4836]: E0122 09:59:08.111099 4836 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-wxfgn" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="registry-server" Jan 22 09:59:08 crc kubenswrapper[4836]: I0122 09:59:08.883829 4836 generic.go:334] "Generic (PLEG): container finished" podID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerID="c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" exitCode=0 Jan 22 09:59:08 crc kubenswrapper[4836]: I0122 09:59:08.884103 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerDied","Data":"c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91"} Jan 22 09:59:08 crc kubenswrapper[4836]: I0122 09:59:08.945555 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.084714 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp\") pod \"b79caeee-9bca-44cf-89ac-520e9b0732cc\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.084879 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content\") pod \"b79caeee-9bca-44cf-89ac-520e9b0732cc\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.084904 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities\") pod \"b79caeee-9bca-44cf-89ac-520e9b0732cc\" (UID: \"b79caeee-9bca-44cf-89ac-520e9b0732cc\") " Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.085870 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities" (OuterVolumeSpecName: "utilities") pod "b79caeee-9bca-44cf-89ac-520e9b0732cc" (UID: "b79caeee-9bca-44cf-89ac-520e9b0732cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.135504 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b79caeee-9bca-44cf-89ac-520e9b0732cc" (UID: "b79caeee-9bca-44cf-89ac-520e9b0732cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.150955 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp" (OuterVolumeSpecName: "kube-api-access-r59mp") pod "b79caeee-9bca-44cf-89ac-520e9b0732cc" (UID: "b79caeee-9bca-44cf-89ac-520e9b0732cc"). InnerVolumeSpecName "kube-api-access-r59mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.186679 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r59mp\" (UniqueName: \"kubernetes.io/projected/b79caeee-9bca-44cf-89ac-520e9b0732cc-kube-api-access-r59mp\") on node \"crc\" DevicePath \"\"" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.187003 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.187115 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b79caeee-9bca-44cf-89ac-520e9b0732cc-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.892700 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wxfgn" event={"ID":"b79caeee-9bca-44cf-89ac-520e9b0732cc","Type":"ContainerDied","Data":"2bf91bd1710e71b2f385c64dd3be3911589e45424657645c49635556cd70b59d"} Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.893097 4836 scope.go:117] "RemoveContainer" containerID="c64fd20bcb3f384daf10f45ab842c927ba6a0f2b82b4f8d7dbf1721723f17c91" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.892736 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wxfgn" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.924779 4836 scope.go:117] "RemoveContainer" containerID="2f82c9addd08a4cbdb2162c8655b28fefed9649efa2fd4c6a6447aa90ef42385" Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.938433 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.943691 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wxfgn"] Jan 22 09:59:09 crc kubenswrapper[4836]: I0122 09:59:09.944205 4836 scope.go:117] "RemoveContainer" containerID="2ee5df5f7b69fcd7e663472bedd6b7f581de1db615ad2493e0bdd1ccbf522342" Jan 22 09:59:11 crc kubenswrapper[4836]: I0122 09:59:11.004212 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" path="/var/lib/kubelet/pods/b79caeee-9bca-44cf-89ac-520e9b0732cc/volumes" Jan 22 09:59:16 crc kubenswrapper[4836]: I0122 09:59:16.813500 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7fd664ffb8-qplwn" Jan 22 09:59:23 crc kubenswrapper[4836]: I0122 09:59:23.897938 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 09:59:23 crc kubenswrapper[4836]: I0122 09:59:23.898483 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 09:59:23 crc kubenswrapper[4836]: I0122 09:59:23.898533 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 09:59:23 crc kubenswrapper[4836]: I0122 09:59:23.899126 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 09:59:23 crc kubenswrapper[4836]: I0122 09:59:23.899179 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7" gracePeriod=600 Jan 22 09:59:25 crc kubenswrapper[4836]: I0122 09:59:25.006176 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7" exitCode=0 Jan 22 09:59:25 crc kubenswrapper[4836]: I0122 09:59:25.006258 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7"} Jan 22 09:59:25 crc kubenswrapper[4836]: I0122 09:59:25.006777 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067"} Jan 22 09:59:25 crc kubenswrapper[4836]: I0122 09:59:25.006803 4836 scope.go:117] "RemoveContainer" containerID="04f16e59582182eef0487d89a65a38501dc47fea3f66cb6ac636b97107873dfd" Jan 22 09:59:36 crc kubenswrapper[4836]: I0122 09:59:36.562919 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-799d7d6dcc-bvjvp" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.307995 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59"] Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308638 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="extract-utilities" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308658 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="extract-utilities" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308675 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308683 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308697 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="extract-content" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308705 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="extract-content" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308724 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="extract-content" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308732 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="extract-content" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308743 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308750 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.308761 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="extract-utilities" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308768 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="extract-utilities" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308893 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79caeee-9bca-44cf-89ac-520e9b0732cc" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.308916 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ace0d4-386b-486d-bef4-2a68a812d7a9" containerName="registry-server" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.309432 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.311504 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.316371 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pvr6q"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.316871 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-pbld2" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.322489 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.324828 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.325171 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.326956 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.339229 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-metrics\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.340771 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-reloader\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.340810 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.340849 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-conf\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.340994 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-sockets\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.341050 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f3969312-f604-49ff-9560-a6078e253ee9-frr-startup\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.341092 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfdvx\" (UniqueName: \"kubernetes.io/projected/c11f6970-4b6e-4574-b9a6-0d50617198dc-kube-api-access-jfdvx\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.341122 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58sm4\" (UniqueName: \"kubernetes.io/projected/f3969312-f604-49ff-9560-a6078e253ee9-kube-api-access-58sm4\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.341177 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.413693 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-v5n2k"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.414606 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.419557 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.419727 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.419769 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.420030 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-vv9wr" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.435762 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-phjc8"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.436929 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.439076 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445664 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445715 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m58zg\" (UniqueName: \"kubernetes.io/projected/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-kube-api-access-m58zg\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445745 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-metrics\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445768 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metrics-certs\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445797 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-reloader\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445814 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445835 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-conf\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445864 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-sockets\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445878 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f3969312-f604-49ff-9560-a6078e253ee9-frr-startup\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445894 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445916 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metallb-excludel2\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445934 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfdvx\" (UniqueName: \"kubernetes.io/projected/c11f6970-4b6e-4574-b9a6-0d50617198dc-kube-api-access-jfdvx\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.445949 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58sm4\" (UniqueName: \"kubernetes.io/projected/f3969312-f604-49ff-9560-a6078e253ee9-kube-api-access-58sm4\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.446248 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-metrics\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.446443 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-sockets\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.446619 4836 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.446669 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs podName:f3969312-f604-49ff-9560-a6078e253ee9 nodeName:}" failed. No retries permitted until 2026-01-22 09:59:37.946652651 +0000 UTC m=+1061.306421168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs") pod "frr-k8s-pvr6q" (UID: "f3969312-f604-49ff-9560-a6078e253ee9") : secret "frr-k8s-certs-secret" not found Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.446786 4836 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.446923 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert podName:c11f6970-4b6e-4574-b9a6-0d50617198dc nodeName:}" failed. No retries permitted until 2026-01-22 09:59:37.946899708 +0000 UTC m=+1061.306668295 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert") pod "frr-k8s-webhook-server-7df86c4f6c-4lq59" (UID: "c11f6970-4b6e-4574-b9a6-0d50617198dc") : secret "frr-k8s-webhook-server-cert" not found Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.447116 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-reloader\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.447395 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f3969312-f604-49ff-9560-a6078e253ee9-frr-conf\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.447907 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f3969312-f604-49ff-9560-a6078e253ee9-frr-startup\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.453928 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-phjc8"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.483946 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58sm4\" (UniqueName: \"kubernetes.io/projected/f3969312-f604-49ff-9560-a6078e253ee9-kube-api-access-58sm4\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.493340 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfdvx\" (UniqueName: \"kubernetes.io/projected/c11f6970-4b6e-4574-b9a6-0d50617198dc-kube-api-access-jfdvx\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547298 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547356 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-metrics-certs\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547385 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metallb-excludel2\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547447 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m58zg\" (UniqueName: \"kubernetes.io/projected/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-kube-api-access-m58zg\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.547467 4836 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 22 09:59:37 crc kubenswrapper[4836]: E0122 09:59:37.547525 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist podName:e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c nodeName:}" failed. No retries permitted until 2026-01-22 09:59:38.047507026 +0000 UTC m=+1061.407275543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist") pod "speaker-v5n2k" (UID: "e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c") : secret "metallb-memberlist" not found Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547471 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metrics-certs\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547799 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-cert\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.547850 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbxpv\" (UniqueName: \"kubernetes.io/projected/9437c57d-62ec-44b0-bf05-2e42c76e6f49-kube-api-access-pbxpv\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.548409 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metallb-excludel2\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.561111 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-metrics-certs\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.568517 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m58zg\" (UniqueName: \"kubernetes.io/projected/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-kube-api-access-m58zg\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.648786 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-cert\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.648867 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbxpv\" (UniqueName: \"kubernetes.io/projected/9437c57d-62ec-44b0-bf05-2e42c76e6f49-kube-api-access-pbxpv\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.648910 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-metrics-certs\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.650800 4836 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.652522 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-metrics-certs\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.662406 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9437c57d-62ec-44b0-bf05-2e42c76e6f49-cert\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.669466 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbxpv\" (UniqueName: \"kubernetes.io/projected/9437c57d-62ec-44b0-bf05-2e42c76e6f49-kube-api-access-pbxpv\") pod \"controller-6968d8fdc4-phjc8\" (UID: \"9437c57d-62ec-44b0-bf05-2e42c76e6f49\") " pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.763724 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.953905 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.954280 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.954947 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-phjc8"] Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.957946 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3969312-f604-49ff-9560-a6078e253ee9-metrics-certs\") pod \"frr-k8s-pvr6q\" (UID: \"f3969312-f604-49ff-9560-a6078e253ee9\") " pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:37 crc kubenswrapper[4836]: I0122 09:59:37.958158 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c11f6970-4b6e-4574-b9a6-0d50617198dc-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-4lq59\" (UID: \"c11f6970-4b6e-4574-b9a6-0d50617198dc\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:38 crc kubenswrapper[4836]: I0122 09:59:38.055656 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:38 crc kubenswrapper[4836]: E0122 09:59:38.055844 4836 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 22 09:59:38 crc kubenswrapper[4836]: E0122 09:59:38.055929 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist podName:e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c nodeName:}" failed. No retries permitted until 2026-01-22 09:59:39.055912049 +0000 UTC m=+1062.415680566 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist") pod "speaker-v5n2k" (UID: "e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c") : secret "metallb-memberlist" not found Jan 22 09:59:38 crc kubenswrapper[4836]: I0122 09:59:38.082732 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-phjc8" event={"ID":"9437c57d-62ec-44b0-bf05-2e42c76e6f49","Type":"ContainerStarted","Data":"ed3fdf03e634517d61c7a414d4d2322a734affa51c2f2a6f4eb350a60e07258a"} Jan 22 09:59:38 crc kubenswrapper[4836]: I0122 09:59:38.232720 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:38 crc kubenswrapper[4836]: I0122 09:59:38.250308 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:38 crc kubenswrapper[4836]: W0122 09:59:38.426655 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc11f6970_4b6e_4574_b9a6_0d50617198dc.slice/crio-7c2b13a94bde8fe2c1b68ed5ebe53d4739d0cd138c491508b86e4fd2d69d3fa0 WatchSource:0}: Error finding container 7c2b13a94bde8fe2c1b68ed5ebe53d4739d0cd138c491508b86e4fd2d69d3fa0: Status 404 returned error can't find the container with id 7c2b13a94bde8fe2c1b68ed5ebe53d4739d0cd138c491508b86e4fd2d69d3fa0 Jan 22 09:59:38 crc kubenswrapper[4836]: I0122 09:59:38.427621 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59"] Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.069202 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.074911 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c-memberlist\") pod \"speaker-v5n2k\" (UID: \"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c\") " pod="metallb-system/speaker-v5n2k" Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.090845 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" event={"ID":"c11f6970-4b6e-4574-b9a6-0d50617198dc","Type":"ContainerStarted","Data":"7c2b13a94bde8fe2c1b68ed5ebe53d4739d0cd138c491508b86e4fd2d69d3fa0"} Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.092247 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"e146a087615f9b8a321bd57f13b9f133edce11b816d594e5597d025dbe15888f"} Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.094616 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-phjc8" event={"ID":"9437c57d-62ec-44b0-bf05-2e42c76e6f49","Type":"ContainerStarted","Data":"62a1afe75fa91c0fa455f5a3d3a5e22387bb15ca0f9c9b64d4dbae1f504826b4"} Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.094669 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-phjc8" event={"ID":"9437c57d-62ec-44b0-bf05-2e42c76e6f49","Type":"ContainerStarted","Data":"d828b1790774fceff862e933826eb3dfbb2d1b2dd1c34157ec86ced98b713e92"} Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.094784 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.110801 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-phjc8" podStartSLOduration=2.110780802 podStartE2EDuration="2.110780802s" podCreationTimestamp="2026-01-22 09:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:59:39.110267369 +0000 UTC m=+1062.470035886" watchObservedRunningTime="2026-01-22 09:59:39.110780802 +0000 UTC m=+1062.470549319" Jan 22 09:59:39 crc kubenswrapper[4836]: I0122 09:59:39.231013 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-v5n2k" Jan 22 09:59:40 crc kubenswrapper[4836]: I0122 09:59:40.118895 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v5n2k" event={"ID":"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c","Type":"ContainerStarted","Data":"e7d82739a04c8a37c2f96a4beec84472fd665f1ebd3cb8440fccce6b34ac27ca"} Jan 22 09:59:40 crc kubenswrapper[4836]: I0122 09:59:40.119405 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v5n2k" event={"ID":"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c","Type":"ContainerStarted","Data":"8ee7e91c2a4e1dc97e5cd9dcd215d67c111a214db3fc6766a873dc8f84ecde5a"} Jan 22 09:59:41 crc kubenswrapper[4836]: I0122 09:59:41.127337 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-v5n2k" event={"ID":"e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c","Type":"ContainerStarted","Data":"51a74a64dba87c26d091b3a00c4f35b4c96d3a4073e74ba2440ced65fd49ce22"} Jan 22 09:59:41 crc kubenswrapper[4836]: I0122 09:59:41.128583 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-v5n2k" Jan 22 09:59:41 crc kubenswrapper[4836]: I0122 09:59:41.156804 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-v5n2k" podStartSLOduration=4.1567888 podStartE2EDuration="4.1567888s" podCreationTimestamp="2026-01-22 09:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:59:41.153841782 +0000 UTC m=+1064.513610299" watchObservedRunningTime="2026-01-22 09:59:41.1567888 +0000 UTC m=+1064.516557317" Jan 22 09:59:46 crc kubenswrapper[4836]: I0122 09:59:46.167305 4836 generic.go:334] "Generic (PLEG): container finished" podID="f3969312-f604-49ff-9560-a6078e253ee9" containerID="5b913aad9d1087f764fb701bc6b6d51ba3f3fb24bfff176987d8310ecd96f5ca" exitCode=0 Jan 22 09:59:46 crc kubenswrapper[4836]: I0122 09:59:46.167375 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerDied","Data":"5b913aad9d1087f764fb701bc6b6d51ba3f3fb24bfff176987d8310ecd96f5ca"} Jan 22 09:59:46 crc kubenswrapper[4836]: I0122 09:59:46.169620 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" event={"ID":"c11f6970-4b6e-4574-b9a6-0d50617198dc","Type":"ContainerStarted","Data":"8d35bd193fa71055eaaee8ce8626a116145efccae7f684d9e0155b324ea5c3c0"} Jan 22 09:59:46 crc kubenswrapper[4836]: I0122 09:59:46.169797 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:46 crc kubenswrapper[4836]: I0122 09:59:46.207976 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" podStartSLOduration=2.252273385 podStartE2EDuration="9.207960791s" podCreationTimestamp="2026-01-22 09:59:37 +0000 UTC" firstStartedPulling="2026-01-22 09:59:38.429508756 +0000 UTC m=+1061.789277263" lastFinishedPulling="2026-01-22 09:59:45.385196152 +0000 UTC m=+1068.744964669" observedRunningTime="2026-01-22 09:59:46.207250332 +0000 UTC m=+1069.567018849" watchObservedRunningTime="2026-01-22 09:59:46.207960791 +0000 UTC m=+1069.567729308" Jan 22 09:59:47 crc kubenswrapper[4836]: I0122 09:59:47.178702 4836 generic.go:334] "Generic (PLEG): container finished" podID="f3969312-f604-49ff-9560-a6078e253ee9" containerID="2d6cb5c0138a56ef7ee86c920995fa51759d95c353cd9bc388ec4733096ab586" exitCode=0 Jan 22 09:59:47 crc kubenswrapper[4836]: I0122 09:59:47.178758 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerDied","Data":"2d6cb5c0138a56ef7ee86c920995fa51759d95c353cd9bc388ec4733096ab586"} Jan 22 09:59:48 crc kubenswrapper[4836]: I0122 09:59:48.189911 4836 generic.go:334] "Generic (PLEG): container finished" podID="f3969312-f604-49ff-9560-a6078e253ee9" containerID="fde5ce664e3fce5c0b3454e8bdaa8b943faffc98757066eae7f158665e2a8e7f" exitCode=0 Jan 22 09:59:48 crc kubenswrapper[4836]: I0122 09:59:48.189972 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerDied","Data":"fde5ce664e3fce5c0b3454e8bdaa8b943faffc98757066eae7f158665e2a8e7f"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.212184 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"877d2c8491ea18a5af9b5b83fc9c7be5a31a4b75075964ef867f8783a08f8829"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.212504 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"0f427e7e4cfa1909e0654d8cd2f46d53766e27658026eec1f2ea0376c707badb"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.212514 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"aebc20d745b0cf112de58bb47dc7dc0f2842fc9d054dc4e9358f818ac3a8ef24"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.212523 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"3bd4aa5db9c4c9b3f0fa3e1c636a37cd34a18a69e71d473699d2e7f62cf20c12"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.212531 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"f3a68880ae2501df7b4e5cec8c084ed0d70203b3af74d8812cacde6b3ea823d1"} Jan 22 09:59:49 crc kubenswrapper[4836]: I0122 09:59:49.246138 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-v5n2k" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.236655 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pvr6q" event={"ID":"f3969312-f604-49ff-9560-a6078e253ee9","Type":"ContainerStarted","Data":"f774a8c42ae60d8fbfcadb5b66c9f194e84bee6e4632de4bd18a0fe1d0e71a2f"} Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.237098 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.271415 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pvr6q" podStartSLOduration=6.247916274 podStartE2EDuration="13.271394638s" podCreationTimestamp="2026-01-22 09:59:37 +0000 UTC" firstStartedPulling="2026-01-22 09:59:38.378275867 +0000 UTC m=+1061.738044384" lastFinishedPulling="2026-01-22 09:59:45.401754191 +0000 UTC m=+1068.761522748" observedRunningTime="2026-01-22 09:59:50.265512653 +0000 UTC m=+1073.625281170" watchObservedRunningTime="2026-01-22 09:59:50.271394638 +0000 UTC m=+1073.631163155" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.600648 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm"] Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.601903 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.604078 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.609601 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm"] Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.637382 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p57np\" (UniqueName: \"kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.637481 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.637744 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.738597 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.738705 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p57np\" (UniqueName: \"kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.738821 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.739177 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.739286 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.768773 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p57np\" (UniqueName: \"kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:50 crc kubenswrapper[4836]: I0122 09:59:50.926365 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 09:59:51 crc kubenswrapper[4836]: I0122 09:59:51.127955 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm"] Jan 22 09:59:51 crc kubenswrapper[4836]: I0122 09:59:51.245156 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" event={"ID":"3b5f2dd9-4392-4345-9241-b0b3a3cc2936","Type":"ContainerStarted","Data":"8b66f48ecb621e8c12724f4be35c85312ee7b7c05c1fdc801f136690dd429e35"} Jan 22 09:59:52 crc kubenswrapper[4836]: I0122 09:59:52.251799 4836 generic.go:334] "Generic (PLEG): container finished" podID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerID="702a1313664864b4f39448a940d9b3e7ca1f6ed80d3b1fb7b22d6fee018c3bc4" exitCode=0 Jan 22 09:59:52 crc kubenswrapper[4836]: I0122 09:59:52.252113 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" event={"ID":"3b5f2dd9-4392-4345-9241-b0b3a3cc2936","Type":"ContainerDied","Data":"702a1313664864b4f39448a940d9b3e7ca1f6ed80d3b1fb7b22d6fee018c3bc4"} Jan 22 09:59:53 crc kubenswrapper[4836]: I0122 09:59:53.251000 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:53 crc kubenswrapper[4836]: I0122 09:59:53.293943 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pvr6q" Jan 22 09:59:57 crc kubenswrapper[4836]: I0122 09:59:57.768015 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-phjc8" Jan 22 09:59:58 crc kubenswrapper[4836]: I0122 09:59:58.238200 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-4lq59" Jan 22 09:59:58 crc kubenswrapper[4836]: I0122 09:59:58.252852 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pvr6q" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.142309 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf"] Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.143428 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.145115 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.146302 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.153152 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf"] Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.165845 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmp2m\" (UniqueName: \"kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.166018 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.166166 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.267528 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.267802 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.267886 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmp2m\" (UniqueName: \"kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.268515 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.274963 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.286730 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmp2m\" (UniqueName: \"kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m\") pod \"collect-profiles-29484600-4bsdf\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.297357 4836 generic.go:334] "Generic (PLEG): container finished" podID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerID="ac6952b0382b64021d65234b2d07710990fc0b94f1da0f3b3ea2bbcb8161dbb6" exitCode=0 Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.297389 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" event={"ID":"3b5f2dd9-4392-4345-9241-b0b3a3cc2936","Type":"ContainerDied","Data":"ac6952b0382b64021d65234b2d07710990fc0b94f1da0f3b3ea2bbcb8161dbb6"} Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.464666 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:00 crc kubenswrapper[4836]: I0122 10:00:00.869703 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf"] Jan 22 10:00:01 crc kubenswrapper[4836]: I0122 10:00:01.303773 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" event={"ID":"d498665e-4a8c-44fe-88b8-6e3fc019367d","Type":"ContainerStarted","Data":"deab9e5882ef205434af5bcb82bc00af529e148a76cca5674ef5e41b2cd7edc3"} Jan 22 10:00:02 crc kubenswrapper[4836]: I0122 10:00:02.310913 4836 generic.go:334] "Generic (PLEG): container finished" podID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerID="b8f13443393e7373425b6c40f360fd216ccff475d46e85c7ba0447d9df13f6fd" exitCode=0 Jan 22 10:00:02 crc kubenswrapper[4836]: I0122 10:00:02.311003 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" event={"ID":"3b5f2dd9-4392-4345-9241-b0b3a3cc2936","Type":"ContainerDied","Data":"b8f13443393e7373425b6c40f360fd216ccff475d46e85c7ba0447d9df13f6fd"} Jan 22 10:00:02 crc kubenswrapper[4836]: I0122 10:00:02.312871 4836 generic.go:334] "Generic (PLEG): container finished" podID="d498665e-4a8c-44fe-88b8-6e3fc019367d" containerID="345df7f330b3b23e3b9c9cc5314acdf21185263bed1062bb79a591f1cb294e78" exitCode=0 Jan 22 10:00:02 crc kubenswrapper[4836]: I0122 10:00:02.312935 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" event={"ID":"d498665e-4a8c-44fe-88b8-6e3fc019367d","Type":"ContainerDied","Data":"345df7f330b3b23e3b9c9cc5314acdf21185263bed1062bb79a591f1cb294e78"} Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.638312 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.639614 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.713464 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle\") pod \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.713616 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume\") pod \"d498665e-4a8c-44fe-88b8-6e3fc019367d\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.713669 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p57np\" (UniqueName: \"kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np\") pod \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.713722 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util\") pod \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\" (UID: \"3b5f2dd9-4392-4345-9241-b0b3a3cc2936\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.713741 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume\") pod \"d498665e-4a8c-44fe-88b8-6e3fc019367d\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.714456 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume" (OuterVolumeSpecName: "config-volume") pod "d498665e-4a8c-44fe-88b8-6e3fc019367d" (UID: "d498665e-4a8c-44fe-88b8-6e3fc019367d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.714520 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmp2m\" (UniqueName: \"kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m\") pod \"d498665e-4a8c-44fe-88b8-6e3fc019367d\" (UID: \"d498665e-4a8c-44fe-88b8-6e3fc019367d\") " Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.715001 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle" (OuterVolumeSpecName: "bundle") pod "3b5f2dd9-4392-4345-9241-b0b3a3cc2936" (UID: "3b5f2dd9-4392-4345-9241-b0b3a3cc2936"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.715089 4836 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d498665e-4a8c-44fe-88b8-6e3fc019367d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.718906 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np" (OuterVolumeSpecName: "kube-api-access-p57np") pod "3b5f2dd9-4392-4345-9241-b0b3a3cc2936" (UID: "3b5f2dd9-4392-4345-9241-b0b3a3cc2936"). InnerVolumeSpecName "kube-api-access-p57np". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.718978 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d498665e-4a8c-44fe-88b8-6e3fc019367d" (UID: "d498665e-4a8c-44fe-88b8-6e3fc019367d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.720031 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m" (OuterVolumeSpecName: "kube-api-access-tmp2m") pod "d498665e-4a8c-44fe-88b8-6e3fc019367d" (UID: "d498665e-4a8c-44fe-88b8-6e3fc019367d"). InnerVolumeSpecName "kube-api-access-tmp2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.727276 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util" (OuterVolumeSpecName: "util") pod "3b5f2dd9-4392-4345-9241-b0b3a3cc2936" (UID: "3b5f2dd9-4392-4345-9241-b0b3a3cc2936"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.816474 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmp2m\" (UniqueName: \"kubernetes.io/projected/d498665e-4a8c-44fe-88b8-6e3fc019367d-kube-api-access-tmp2m\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.816870 4836 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.816889 4836 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d498665e-4a8c-44fe-88b8-6e3fc019367d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.816901 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p57np\" (UniqueName: \"kubernetes.io/projected/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-kube-api-access-p57np\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:03 crc kubenswrapper[4836]: I0122 10:00:03.816909 4836 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b5f2dd9-4392-4345-9241-b0b3a3cc2936-util\") on node \"crc\" DevicePath \"\"" Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.332786 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" event={"ID":"3b5f2dd9-4392-4345-9241-b0b3a3cc2936","Type":"ContainerDied","Data":"8b66f48ecb621e8c12724f4be35c85312ee7b7c05c1fdc801f136690dd429e35"} Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.333235 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b66f48ecb621e8c12724f4be35c85312ee7b7c05c1fdc801f136690dd429e35" Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.333242 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm" Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.340039 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" event={"ID":"d498665e-4a8c-44fe-88b8-6e3fc019367d","Type":"ContainerDied","Data":"deab9e5882ef205434af5bcb82bc00af529e148a76cca5674ef5e41b2cd7edc3"} Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.340100 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deab9e5882ef205434af5bcb82bc00af529e148a76cca5674ef5e41b2cd7edc3" Jan 22 10:00:04 crc kubenswrapper[4836]: I0122 10:00:04.340151 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-4bsdf" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726002 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7"] Jan 22 10:00:08 crc kubenswrapper[4836]: E0122 10:00:08.726610 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d498665e-4a8c-44fe-88b8-6e3fc019367d" containerName="collect-profiles" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726627 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="d498665e-4a8c-44fe-88b8-6e3fc019367d" containerName="collect-profiles" Jan 22 10:00:08 crc kubenswrapper[4836]: E0122 10:00:08.726639 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="util" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726646 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="util" Jan 22 10:00:08 crc kubenswrapper[4836]: E0122 10:00:08.726665 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="pull" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726672 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="pull" Jan 22 10:00:08 crc kubenswrapper[4836]: E0122 10:00:08.726687 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="extract" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726697 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="extract" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726830 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b5f2dd9-4392-4345-9241-b0b3a3cc2936" containerName="extract" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.726841 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="d498665e-4a8c-44fe-88b8-6e3fc019367d" containerName="collect-profiles" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.727342 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.729957 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.730374 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.730646 4836 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-qw6vh" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.740292 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7"] Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.778337 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwsqc\" (UniqueName: \"kubernetes.io/projected/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-kube-api-access-gwsqc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.778624 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.879663 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.879710 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwsqc\" (UniqueName: \"kubernetes.io/projected/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-kube-api-access-gwsqc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.880123 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:08 crc kubenswrapper[4836]: I0122 10:00:08.898490 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwsqc\" (UniqueName: \"kubernetes.io/projected/8d7bfcde-716b-40d9-9b1d-9ce4708e590f-kube-api-access-gwsqc\") pod \"cert-manager-operator-controller-manager-64cf6dff88-rpfk7\" (UID: \"8d7bfcde-716b-40d9-9b1d-9ce4708e590f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:09 crc kubenswrapper[4836]: I0122 10:00:09.048632 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" Jan 22 10:00:09 crc kubenswrapper[4836]: I0122 10:00:09.262976 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7"] Jan 22 10:00:09 crc kubenswrapper[4836]: W0122 10:00:09.275187 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d7bfcde_716b_40d9_9b1d_9ce4708e590f.slice/crio-f2b4b70c31621e37ba075f15e01e4dc77ea0308bec8dc240dbba79f9b1eeef18 WatchSource:0}: Error finding container f2b4b70c31621e37ba075f15e01e4dc77ea0308bec8dc240dbba79f9b1eeef18: Status 404 returned error can't find the container with id f2b4b70c31621e37ba075f15e01e4dc77ea0308bec8dc240dbba79f9b1eeef18 Jan 22 10:00:09 crc kubenswrapper[4836]: I0122 10:00:09.371153 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" event={"ID":"8d7bfcde-716b-40d9-9b1d-9ce4708e590f","Type":"ContainerStarted","Data":"f2b4b70c31621e37ba075f15e01e4dc77ea0308bec8dc240dbba79f9b1eeef18"} Jan 22 10:00:19 crc kubenswrapper[4836]: I0122 10:00:19.470573 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" event={"ID":"8d7bfcde-716b-40d9-9b1d-9ce4708e590f","Type":"ContainerStarted","Data":"ccc26f35a18cd63832d140785ade22cc44c484697e98da155af26ec19c0fbd14"} Jan 22 10:00:19 crc kubenswrapper[4836]: I0122 10:00:19.491521 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-rpfk7" podStartSLOduration=2.120118654 podStartE2EDuration="11.491501522s" podCreationTimestamp="2026-01-22 10:00:08 +0000 UTC" firstStartedPulling="2026-01-22 10:00:09.280144049 +0000 UTC m=+1092.639912556" lastFinishedPulling="2026-01-22 10:00:18.651526907 +0000 UTC m=+1102.011295424" observedRunningTime="2026-01-22 10:00:19.48612213 +0000 UTC m=+1102.845890657" watchObservedRunningTime="2026-01-22 10:00:19.491501522 +0000 UTC m=+1102.851270039" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.755645 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-b8jpq"] Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.756514 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.758917 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.763075 4836 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-wcchz" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.763103 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.782052 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-b8jpq"] Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.875845 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.875922 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbsp4\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-kube-api-access-kbsp4\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.977593 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:23 crc kubenswrapper[4836]: I0122 10:00:23.977648 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbsp4\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-kube-api-access-kbsp4\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.006957 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbsp4\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-kube-api-access-kbsp4\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.007368 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b47c9aca-4964-4a7f-8588-622b55539ec8-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-b8jpq\" (UID: \"b47c9aca-4964-4a7f-8588-622b55539ec8\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.075686 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.413653 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-b8jpq"] Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.500433 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" event={"ID":"b47c9aca-4964-4a7f-8588-622b55539ec8","Type":"ContainerStarted","Data":"52c223d548d8cdf64b1c1689075e1b96fb0fd882dd5898e03d6cda3d418c6037"} Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.982327 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c"] Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.983681 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.989348 4836 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vzffb" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.996150 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c"] Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.998990 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:24 crc kubenswrapper[4836]: I0122 10:00:24.999123 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wf8n\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-kube-api-access-5wf8n\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.100231 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.100336 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wf8n\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-kube-api-access-5wf8n\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.117661 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.117761 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wf8n\" (UniqueName: \"kubernetes.io/projected/9994d233-12a4-49af-8e54-b76382ca5efb-kube-api-access-5wf8n\") pod \"cert-manager-cainjector-855d9ccff4-jhm6c\" (UID: \"9994d233-12a4-49af-8e54-b76382ca5efb\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.307828 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" Jan 22 10:00:25 crc kubenswrapper[4836]: I0122 10:00:25.753959 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c"] Jan 22 10:00:26 crc kubenswrapper[4836]: I0122 10:00:26.515061 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" event={"ID":"9994d233-12a4-49af-8e54-b76382ca5efb","Type":"ContainerStarted","Data":"51057dd542f5e83d578b1d67d207c1c51efbd7fd1c376a7d41a23eaedd830581"} Jan 22 10:00:39 crc kubenswrapper[4836]: E0122 10:00:39.509260 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df" Jan 22 10:00:39 crc kubenswrapper[4836]: E0122 10:00:39.509821 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-webhook,Image:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,Command:[/app/cmd/webhook/webhook],Args:[--dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.$(POD_NAMESPACE),cert-manager-webhook.$(POD_NAMESPACE).svc --secure-port=10250 --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:10250,Protocol:TCP,HostIP:,},ContainerPort{Name:healthcheck,HostPort:0,ContainerPort:6080,Protocol:TCP,HostIP:,},ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bound-sa-token,ReadOnly:true,MountPath:/var/run/secrets/openshift/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbsp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:60,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000710000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-webhook-f4fb5df64-b8jpq_cert-manager(b47c9aca-4964-4a7f-8588-622b55539ec8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 22 10:00:39 crc kubenswrapper[4836]: E0122 10:00:39.511459 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" podUID="b47c9aca-4964-4a7f-8588-622b55539ec8" Jan 22 10:00:40 crc kubenswrapper[4836]: I0122 10:00:40.607967 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" event={"ID":"b47c9aca-4964-4a7f-8588-622b55539ec8","Type":"ContainerStarted","Data":"83a026ae4785196f5fbeff6fc87521b30074d8484bedf560f2225e6f6831e9ac"} Jan 22 10:00:40 crc kubenswrapper[4836]: I0122 10:00:40.609282 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:40 crc kubenswrapper[4836]: I0122 10:00:40.609696 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" event={"ID":"9994d233-12a4-49af-8e54-b76382ca5efb","Type":"ContainerStarted","Data":"e519d20849c52696bccecfb43c470a2faca76d2dce91ed51a5d7066d2cb49377"} Jan 22 10:00:40 crc kubenswrapper[4836]: I0122 10:00:40.623985 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" podStartSLOduration=-9223372019.230806 podStartE2EDuration="17.623969035s" podCreationTimestamp="2026-01-22 10:00:23 +0000 UTC" firstStartedPulling="2026-01-22 10:00:24.422727884 +0000 UTC m=+1107.782496401" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:00:40.62303088 +0000 UTC m=+1123.982799397" watchObservedRunningTime="2026-01-22 10:00:40.623969035 +0000 UTC m=+1123.983737552" Jan 22 10:00:40 crc kubenswrapper[4836]: I0122 10:00:40.638502 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jhm6c" podStartSLOduration=2.470044068 podStartE2EDuration="16.638487058s" podCreationTimestamp="2026-01-22 10:00:24 +0000 UTC" firstStartedPulling="2026-01-22 10:00:25.783862631 +0000 UTC m=+1109.143631158" lastFinishedPulling="2026-01-22 10:00:39.952305631 +0000 UTC m=+1123.312074148" observedRunningTime="2026-01-22 10:00:40.63478317 +0000 UTC m=+1123.994551687" watchObservedRunningTime="2026-01-22 10:00:40.638487058 +0000 UTC m=+1123.998255575" Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.893119 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4qvrz"] Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.894271 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.897212 4836 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7gcsw" Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.912479 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4qvrz"] Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.959970 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcwjl\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-kube-api-access-kcwjl\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:41 crc kubenswrapper[4836]: I0122 10:00:41.960090 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-bound-sa-token\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.061885 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcwjl\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-kube-api-access-kcwjl\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.062338 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-bound-sa-token\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.093022 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-bound-sa-token\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.094553 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcwjl\" (UniqueName: \"kubernetes.io/projected/dde5f84b-448c-4059-9adb-e5858090d8c4-kube-api-access-kcwjl\") pod \"cert-manager-86cb77c54b-4qvrz\" (UID: \"dde5f84b-448c-4059-9adb-e5858090d8c4\") " pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.228598 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4qvrz" Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.476491 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4qvrz"] Jan 22 10:00:42 crc kubenswrapper[4836]: W0122 10:00:42.482769 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddde5f84b_448c_4059_9adb_e5858090d8c4.slice/crio-87ca95536d9c9f6112bceb8951e3b388d8c5d8e0f867246171115061e8d48a16 WatchSource:0}: Error finding container 87ca95536d9c9f6112bceb8951e3b388d8c5d8e0f867246171115061e8d48a16: Status 404 returned error can't find the container with id 87ca95536d9c9f6112bceb8951e3b388d8c5d8e0f867246171115061e8d48a16 Jan 22 10:00:42 crc kubenswrapper[4836]: I0122 10:00:42.623609 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4qvrz" event={"ID":"dde5f84b-448c-4059-9adb-e5858090d8c4","Type":"ContainerStarted","Data":"87ca95536d9c9f6112bceb8951e3b388d8c5d8e0f867246171115061e8d48a16"} Jan 22 10:00:43 crc kubenswrapper[4836]: I0122 10:00:43.633134 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4qvrz" event={"ID":"dde5f84b-448c-4059-9adb-e5858090d8c4","Type":"ContainerStarted","Data":"8adc95b2198428d3f1b9392fdd67fc94f0532b9c93ce7428f6f14372a5888ede"} Jan 22 10:00:43 crc kubenswrapper[4836]: I0122 10:00:43.655018 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-4qvrz" podStartSLOduration=2.6549984330000003 podStartE2EDuration="2.654998433s" podCreationTimestamp="2026-01-22 10:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:00:43.653081223 +0000 UTC m=+1127.012849740" watchObservedRunningTime="2026-01-22 10:00:43.654998433 +0000 UTC m=+1127.014766970" Jan 22 10:00:49 crc kubenswrapper[4836]: I0122 10:00:49.079088 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-b8jpq" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.567102 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.568327 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.574918 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.577315 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vswmh" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.578503 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.582204 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.621679 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxhgq\" (UniqueName: \"kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq\") pod \"openstack-operator-index-g6lzj\" (UID: \"be3dffd6-66ee-47dc-bfbc-e610850add4e\") " pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.722859 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxhgq\" (UniqueName: \"kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq\") pod \"openstack-operator-index-g6lzj\" (UID: \"be3dffd6-66ee-47dc-bfbc-e610850add4e\") " pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.746388 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxhgq\" (UniqueName: \"kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq\") pod \"openstack-operator-index-g6lzj\" (UID: \"be3dffd6-66ee-47dc-bfbc-e610850add4e\") " pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:00:52 crc kubenswrapper[4836]: I0122 10:00:52.886580 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:00:53 crc kubenswrapper[4836]: I0122 10:00:53.366460 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:00:53 crc kubenswrapper[4836]: W0122 10:00:53.371997 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe3dffd6_66ee_47dc_bfbc_e610850add4e.slice/crio-3fc7af5357ec6fa34cd40c8a28925b426424059f4b6168364ffb29bcc6c79651 WatchSource:0}: Error finding container 3fc7af5357ec6fa34cd40c8a28925b426424059f4b6168364ffb29bcc6c79651: Status 404 returned error can't find the container with id 3fc7af5357ec6fa34cd40c8a28925b426424059f4b6168364ffb29bcc6c79651 Jan 22 10:00:53 crc kubenswrapper[4836]: I0122 10:00:53.703996 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g6lzj" event={"ID":"be3dffd6-66ee-47dc-bfbc-e610850add4e","Type":"ContainerStarted","Data":"3fc7af5357ec6fa34cd40c8a28925b426424059f4b6168364ffb29bcc6c79651"} Jan 22 10:00:54 crc kubenswrapper[4836]: I0122 10:00:54.540357 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.159024 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vsgjh"] Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.160471 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.165798 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vsgjh"] Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.274659 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmr6v\" (UniqueName: \"kubernetes.io/projected/46e9eb65-0aaa-45dc-810a-7c0e8cc089a5-kube-api-access-pmr6v\") pod \"openstack-operator-index-vsgjh\" (UID: \"46e9eb65-0aaa-45dc-810a-7c0e8cc089a5\") " pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.375955 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmr6v\" (UniqueName: \"kubernetes.io/projected/46e9eb65-0aaa-45dc-810a-7c0e8cc089a5-kube-api-access-pmr6v\") pod \"openstack-operator-index-vsgjh\" (UID: \"46e9eb65-0aaa-45dc-810a-7c0e8cc089a5\") " pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.412033 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmr6v\" (UniqueName: \"kubernetes.io/projected/46e9eb65-0aaa-45dc-810a-7c0e8cc089a5-kube-api-access-pmr6v\") pod \"openstack-operator-index-vsgjh\" (UID: \"46e9eb65-0aaa-45dc-810a-7c0e8cc089a5\") " pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.486462 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:00:55 crc kubenswrapper[4836]: I0122 10:00:55.919951 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vsgjh"] Jan 22 10:00:55 crc kubenswrapper[4836]: W0122 10:00:55.927379 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46e9eb65_0aaa_45dc_810a_7c0e8cc089a5.slice/crio-832a21f69e986740ad534d79de8647d4ef6873bc8baace18eaf485b3c3656984 WatchSource:0}: Error finding container 832a21f69e986740ad534d79de8647d4ef6873bc8baace18eaf485b3c3656984: Status 404 returned error can't find the container with id 832a21f69e986740ad534d79de8647d4ef6873bc8baace18eaf485b3c3656984 Jan 22 10:00:56 crc kubenswrapper[4836]: I0122 10:00:56.723811 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vsgjh" event={"ID":"46e9eb65-0aaa-45dc-810a-7c0e8cc089a5","Type":"ContainerStarted","Data":"832a21f69e986740ad534d79de8647d4ef6873bc8baace18eaf485b3c3656984"} Jan 22 10:01:00 crc kubenswrapper[4836]: I0122 10:01:00.754579 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g6lzj" event={"ID":"be3dffd6-66ee-47dc-bfbc-e610850add4e","Type":"ContainerStarted","Data":"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa"} Jan 22 10:01:00 crc kubenswrapper[4836]: I0122 10:01:00.754702 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-g6lzj" podUID="be3dffd6-66ee-47dc-bfbc-e610850add4e" containerName="registry-server" containerID="cri-o://fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa" gracePeriod=2 Jan 22 10:01:00 crc kubenswrapper[4836]: I0122 10:01:00.756326 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vsgjh" event={"ID":"46e9eb65-0aaa-45dc-810a-7c0e8cc089a5","Type":"ContainerStarted","Data":"e7d73d990af8227e959d21401b34b1ebb14834e0eb946ba7b19599cc87fd9117"} Jan 22 10:01:00 crc kubenswrapper[4836]: I0122 10:01:00.777766 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-g6lzj" podStartSLOduration=2.131359091 podStartE2EDuration="8.777741301s" podCreationTimestamp="2026-01-22 10:00:52 +0000 UTC" firstStartedPulling="2026-01-22 10:00:53.374083851 +0000 UTC m=+1136.733852368" lastFinishedPulling="2026-01-22 10:01:00.020466051 +0000 UTC m=+1143.380234578" observedRunningTime="2026-01-22 10:01:00.771124116 +0000 UTC m=+1144.130892693" watchObservedRunningTime="2026-01-22 10:01:00.777741301 +0000 UTC m=+1144.137509828" Jan 22 10:01:00 crc kubenswrapper[4836]: I0122 10:01:00.797870 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vsgjh" podStartSLOduration=1.703334525 podStartE2EDuration="5.7978413s" podCreationTimestamp="2026-01-22 10:00:55 +0000 UTC" firstStartedPulling="2026-01-22 10:00:55.930112036 +0000 UTC m=+1139.289880553" lastFinishedPulling="2026-01-22 10:01:00.024618811 +0000 UTC m=+1143.384387328" observedRunningTime="2026-01-22 10:01:00.787419445 +0000 UTC m=+1144.147187972" watchObservedRunningTime="2026-01-22 10:01:00.7978413 +0000 UTC m=+1144.157609837" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.163303 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.260996 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxhgq\" (UniqueName: \"kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq\") pod \"be3dffd6-66ee-47dc-bfbc-e610850add4e\" (UID: \"be3dffd6-66ee-47dc-bfbc-e610850add4e\") " Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.265949 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq" (OuterVolumeSpecName: "kube-api-access-xxhgq") pod "be3dffd6-66ee-47dc-bfbc-e610850add4e" (UID: "be3dffd6-66ee-47dc-bfbc-e610850add4e"). InnerVolumeSpecName "kube-api-access-xxhgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.362814 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxhgq\" (UniqueName: \"kubernetes.io/projected/be3dffd6-66ee-47dc-bfbc-e610850add4e-kube-api-access-xxhgq\") on node \"crc\" DevicePath \"\"" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.765206 4836 generic.go:334] "Generic (PLEG): container finished" podID="be3dffd6-66ee-47dc-bfbc-e610850add4e" containerID="fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa" exitCode=0 Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.765252 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g6lzj" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.765305 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g6lzj" event={"ID":"be3dffd6-66ee-47dc-bfbc-e610850add4e","Type":"ContainerDied","Data":"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa"} Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.765353 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g6lzj" event={"ID":"be3dffd6-66ee-47dc-bfbc-e610850add4e","Type":"ContainerDied","Data":"3fc7af5357ec6fa34cd40c8a28925b426424059f4b6168364ffb29bcc6c79651"} Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.765373 4836 scope.go:117] "RemoveContainer" containerID="fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.785227 4836 scope.go:117] "RemoveContainer" containerID="fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa" Jan 22 10:01:01 crc kubenswrapper[4836]: E0122 10:01:01.785655 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa\": container with ID starting with fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa not found: ID does not exist" containerID="fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.785693 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa"} err="failed to get container status \"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa\": rpc error: code = NotFound desc = could not find container \"fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa\": container with ID starting with fd6123886d8f1d0cc2449a89af45a164fcd9e4f58d4f146b93046f3cd97c85fa not found: ID does not exist" Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.801578 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:01:01 crc kubenswrapper[4836]: I0122 10:01:01.810444 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-g6lzj"] Jan 22 10:01:03 crc kubenswrapper[4836]: I0122 10:01:03.008979 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be3dffd6-66ee-47dc-bfbc-e610850add4e" path="/var/lib/kubelet/pods/be3dffd6-66ee-47dc-bfbc-e610850add4e/volumes" Jan 22 10:01:05 crc kubenswrapper[4836]: I0122 10:01:05.487838 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:01:05 crc kubenswrapper[4836]: I0122 10:01:05.487912 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:01:05 crc kubenswrapper[4836]: I0122 10:01:05.516510 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:01:05 crc kubenswrapper[4836]: I0122 10:01:05.819468 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vsgjh" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.410303 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m"] Jan 22 10:01:13 crc kubenswrapper[4836]: E0122 10:01:13.411332 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3dffd6-66ee-47dc-bfbc-e610850add4e" containerName="registry-server" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.411350 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3dffd6-66ee-47dc-bfbc-e610850add4e" containerName="registry-server" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.411485 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3dffd6-66ee-47dc-bfbc-e610850add4e" containerName="registry-server" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.412686 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.420714 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hdx9x" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.434786 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m"] Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.545368 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.545425 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.545526 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67964\" (UniqueName: \"kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.646884 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67964\" (UniqueName: \"kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.646991 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.647029 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.647633 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.647708 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.670232 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67964\" (UniqueName: \"kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:13 crc kubenswrapper[4836]: I0122 10:01:13.747529 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:14 crc kubenswrapper[4836]: I0122 10:01:14.011823 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m"] Jan 22 10:01:14 crc kubenswrapper[4836]: I0122 10:01:14.860700 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" event={"ID":"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8","Type":"ContainerStarted","Data":"bec4f8ce64b42d39745e8b4191c0492a26d7eb06da4339d4114f12a0d2eb98c3"} Jan 22 10:01:15 crc kubenswrapper[4836]: I0122 10:01:15.872979 4836 generic.go:334] "Generic (PLEG): container finished" podID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerID="0836611c0f9c66c5dd2718f428dc2711d63a11f40bc7c88e5fa2bed8ef166886" exitCode=0 Jan 22 10:01:15 crc kubenswrapper[4836]: I0122 10:01:15.873229 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" event={"ID":"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8","Type":"ContainerDied","Data":"0836611c0f9c66c5dd2718f428dc2711d63a11f40bc7c88e5fa2bed8ef166886"} Jan 22 10:01:16 crc kubenswrapper[4836]: I0122 10:01:16.881229 4836 generic.go:334] "Generic (PLEG): container finished" podID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerID="9ea8728640c442583043658f515ef4d76a1d3d01eaae8661f465d7447b4d83c5" exitCode=0 Jan 22 10:01:16 crc kubenswrapper[4836]: I0122 10:01:16.881290 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" event={"ID":"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8","Type":"ContainerDied","Data":"9ea8728640c442583043658f515ef4d76a1d3d01eaae8661f465d7447b4d83c5"} Jan 22 10:01:17 crc kubenswrapper[4836]: I0122 10:01:17.895900 4836 generic.go:334] "Generic (PLEG): container finished" podID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerID="9842594a680746c86579406ae7760ef2bfffdb96691a45f18809995673159f08" exitCode=0 Jan 22 10:01:17 crc kubenswrapper[4836]: I0122 10:01:17.896020 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" event={"ID":"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8","Type":"ContainerDied","Data":"9842594a680746c86579406ae7760ef2bfffdb96691a45f18809995673159f08"} Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.202641 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.329146 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util\") pod \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.329304 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67964\" (UniqueName: \"kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964\") pod \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.329439 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle\") pod \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\" (UID: \"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8\") " Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.330906 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle" (OuterVolumeSpecName: "bundle") pod "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" (UID: "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.339610 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964" (OuterVolumeSpecName: "kube-api-access-67964") pod "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" (UID: "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8"). InnerVolumeSpecName "kube-api-access-67964". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.363731 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util" (OuterVolumeSpecName: "util") pod "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" (UID: "76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.431367 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67964\" (UniqueName: \"kubernetes.io/projected/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-kube-api-access-67964\") on node \"crc\" DevicePath \"\"" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.431418 4836 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-bundle\") on node \"crc\" DevicePath \"\"" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.431431 4836 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8-util\") on node \"crc\" DevicePath \"\"" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.915658 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" event={"ID":"76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8","Type":"ContainerDied","Data":"bec4f8ce64b42d39745e8b4191c0492a26d7eb06da4339d4114f12a0d2eb98c3"} Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.915733 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bec4f8ce64b42d39745e8b4191c0492a26d7eb06da4339d4114f12a0d2eb98c3" Jan 22 10:01:19 crc kubenswrapper[4836]: I0122 10:01:19.915836 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.793250 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj"] Jan 22 10:01:26 crc kubenswrapper[4836]: E0122 10:01:26.793959 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="pull" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.793973 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="pull" Jan 22 10:01:26 crc kubenswrapper[4836]: E0122 10:01:26.793990 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="util" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.793998 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="util" Jan 22 10:01:26 crc kubenswrapper[4836]: E0122 10:01:26.794012 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="extract" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.794020 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="extract" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.794181 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8" containerName="extract" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.794634 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.797012 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-zfzk5" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.825543 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj"] Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.844389 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sfzl\" (UniqueName: \"kubernetes.io/projected/1782e8eb-b211-494d-8091-45ef1630b6c6-kube-api-access-9sfzl\") pod \"openstack-operator-controller-init-698d6bb84b-69wfj\" (UID: \"1782e8eb-b211-494d-8091-45ef1630b6c6\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.946280 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sfzl\" (UniqueName: \"kubernetes.io/projected/1782e8eb-b211-494d-8091-45ef1630b6c6-kube-api-access-9sfzl\") pod \"openstack-operator-controller-init-698d6bb84b-69wfj\" (UID: \"1782e8eb-b211-494d-8091-45ef1630b6c6\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:26 crc kubenswrapper[4836]: I0122 10:01:26.964157 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sfzl\" (UniqueName: \"kubernetes.io/projected/1782e8eb-b211-494d-8091-45ef1630b6c6-kube-api-access-9sfzl\") pod \"openstack-operator-controller-init-698d6bb84b-69wfj\" (UID: \"1782e8eb-b211-494d-8091-45ef1630b6c6\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:27 crc kubenswrapper[4836]: I0122 10:01:27.116584 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:27 crc kubenswrapper[4836]: I0122 10:01:27.352262 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj"] Jan 22 10:01:27 crc kubenswrapper[4836]: I0122 10:01:27.975257 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" event={"ID":"1782e8eb-b211-494d-8091-45ef1630b6c6","Type":"ContainerStarted","Data":"223d24908be07360f1882d4ed6dbfd19685c1b833eaede5f44cf57b3e0757e6b"} Jan 22 10:01:39 crc kubenswrapper[4836]: I0122 10:01:39.074397 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" event={"ID":"1782e8eb-b211-494d-8091-45ef1630b6c6","Type":"ContainerStarted","Data":"f54b93fd9f706633a7188e8ee16dd9f9a1d1bab818c7d552558b8e3c6694e3bf"} Jan 22 10:01:39 crc kubenswrapper[4836]: I0122 10:01:39.075009 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:39 crc kubenswrapper[4836]: I0122 10:01:39.110309 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" podStartSLOduration=2.126296621 podStartE2EDuration="13.110279668s" podCreationTimestamp="2026-01-22 10:01:26 +0000 UTC" firstStartedPulling="2026-01-22 10:01:27.359493446 +0000 UTC m=+1170.719261963" lastFinishedPulling="2026-01-22 10:01:38.343476493 +0000 UTC m=+1181.703245010" observedRunningTime="2026-01-22 10:01:39.102141648 +0000 UTC m=+1182.461910165" watchObservedRunningTime="2026-01-22 10:01:39.110279668 +0000 UTC m=+1182.470048225" Jan 22 10:01:47 crc kubenswrapper[4836]: I0122 10:01:47.119779 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-69wfj" Jan 22 10:01:54 crc kubenswrapper[4836]: I0122 10:01:53.898111 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:01:54 crc kubenswrapper[4836]: I0122 10:01:53.898438 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.867286 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.868577 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.870203 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5hk65" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.924564 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2855\" (UniqueName: \"kubernetes.io/projected/bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b-kube-api-access-p2855\") pod \"barbican-operator-controller-manager-59dd8b7cbf-7dnbq\" (UID: \"bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.928196 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.929013 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.933390 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.934234 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.936455 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-f8p98" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.936655 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-46nqt" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.939904 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.945922 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.954113 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.979441 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.980138 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.983225 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zzjn4" Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.991879 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.997991 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h"] Jan 22 10:02:13 crc kubenswrapper[4836]: I0122 10:02:13.999056 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.006613 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.009287 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-bl88w" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.013544 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.014946 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.017377 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tzfq9" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.025291 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2855\" (UniqueName: \"kubernetes.io/projected/bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b-kube-api-access-p2855\") pod \"barbican-operator-controller-manager-59dd8b7cbf-7dnbq\" (UID: \"bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.027769 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.028574 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.031084 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.031429 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-tx5g6" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.035105 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.054606 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.058301 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.069201 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-r7zw2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.070762 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2855\" (UniqueName: \"kubernetes.io/projected/bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b-kube-api-access-p2855\") pod \"barbican-operator-controller-manager-59dd8b7cbf-7dnbq\" (UID: \"bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.095509 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.114818 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.118524 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.119335 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.121801 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2zs2g" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.124963 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.127569 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130094 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvbhc\" (UniqueName: \"kubernetes.io/projected/e5ac1213-b336-49f2-a1c3-93fbbbdc724c-kube-api-access-hvbhc\") pod \"cinder-operator-controller-manager-69cf5d4557-dlkpl\" (UID: \"e5ac1213-b336-49f2-a1c3-93fbbbdc724c\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130319 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5pds\" (UniqueName: \"kubernetes.io/projected/2d3d18eb-0f1a-4552-91d3-6c6bec689f12-kube-api-access-k5pds\") pod \"heat-operator-controller-manager-594c8c9d5d-glr2h\" (UID: \"2d3d18eb-0f1a-4552-91d3-6c6bec689f12\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130358 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130382 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgb8j\" (UniqueName: \"kubernetes.io/projected/45ead174-1e73-410e-a2af-5db3a287e6ef-kube-api-access-pgb8j\") pod \"glance-operator-controller-manager-78fdd796fd-zxzq5\" (UID: \"45ead174-1e73-410e-a2af-5db3a287e6ef\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130410 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9c8n\" (UniqueName: \"kubernetes.io/projected/62432f9a-3dc6-4f17-844e-fd5a6e983097-kube-api-access-z9c8n\") pod \"horizon-operator-controller-manager-77d5c5b54f-hfd9z\" (UID: \"62432f9a-3dc6-4f17-844e-fd5a6e983097\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130430 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7x85\" (UniqueName: \"kubernetes.io/projected/43c5f5ad-e7e7-460d-9ed2-555cc840253f-kube-api-access-b7x85\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130454 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt4z2\" (UniqueName: \"kubernetes.io/projected/84ad66e0-8ccb-4bd1-a5c6-ec82697798ed-kube-api-access-bt4z2\") pod \"keystone-operator-controller-manager-b8b6d4659-vf8kq\" (UID: \"84ad66e0-8ccb-4bd1-a5c6-ec82697798ed\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130481 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qggxk\" (UniqueName: \"kubernetes.io/projected/4f4c9bb1-0e84-4d29-9da3-c068c13723db-kube-api-access-qggxk\") pod \"designate-operator-controller-manager-b45d7bf98-hpfrj\" (UID: \"4f4c9bb1-0e84-4d29-9da3-c068c13723db\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130523 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxw59\" (UniqueName: \"kubernetes.io/projected/4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf-kube-api-access-bxw59\") pod \"manila-operator-controller-manager-78c6999f6f-h5xn2\" (UID: \"4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.130576 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhnvb\" (UniqueName: \"kubernetes.io/projected/72203b24-67d7-4dac-8430-51c940eaa7e8-kube-api-access-hhnvb\") pod \"ironic-operator-controller-manager-69d6c9f5b8-gh4kf\" (UID: \"72203b24-67d7-4dac-8430-51c940eaa7e8\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.133036 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-fmzvd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.160276 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.186419 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.194550 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.195445 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.204285 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-xglpd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.205133 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.208243 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.216412 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ql655" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.227436 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.233651 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5pds\" (UniqueName: \"kubernetes.io/projected/2d3d18eb-0f1a-4552-91d3-6c6bec689f12-kube-api-access-k5pds\") pod \"heat-operator-controller-manager-594c8c9d5d-glr2h\" (UID: \"2d3d18eb-0f1a-4552-91d3-6c6bec689f12\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.239639 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgb8j\" (UniqueName: \"kubernetes.io/projected/45ead174-1e73-410e-a2af-5db3a287e6ef-kube-api-access-pgb8j\") pod \"glance-operator-controller-manager-78fdd796fd-zxzq5\" (UID: \"45ead174-1e73-410e-a2af-5db3a287e6ef\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.239664 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.239704 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q44t5\" (UniqueName: \"kubernetes.io/projected/f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7-kube-api-access-q44t5\") pod \"neutron-operator-controller-manager-5d8f59fb49-9v24b\" (UID: \"f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.239741 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9c8n\" (UniqueName: \"kubernetes.io/projected/62432f9a-3dc6-4f17-844e-fd5a6e983097-kube-api-access-z9c8n\") pod \"horizon-operator-controller-manager-77d5c5b54f-hfd9z\" (UID: \"62432f9a-3dc6-4f17-844e-fd5a6e983097\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240182 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7x85\" (UniqueName: \"kubernetes.io/projected/43c5f5ad-e7e7-460d-9ed2-555cc840253f-kube-api-access-b7x85\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240249 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt4z2\" (UniqueName: \"kubernetes.io/projected/84ad66e0-8ccb-4bd1-a5c6-ec82697798ed-kube-api-access-bt4z2\") pod \"keystone-operator-controller-manager-b8b6d4659-vf8kq\" (UID: \"84ad66e0-8ccb-4bd1-a5c6-ec82697798ed\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240285 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qggxk\" (UniqueName: \"kubernetes.io/projected/4f4c9bb1-0e84-4d29-9da3-c068c13723db-kube-api-access-qggxk\") pod \"designate-operator-controller-manager-b45d7bf98-hpfrj\" (UID: \"4f4c9bb1-0e84-4d29-9da3-c068c13723db\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240347 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxw59\" (UniqueName: \"kubernetes.io/projected/4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf-kube-api-access-bxw59\") pod \"manila-operator-controller-manager-78c6999f6f-h5xn2\" (UID: \"4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240439 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhnvb\" (UniqueName: \"kubernetes.io/projected/72203b24-67d7-4dac-8430-51c940eaa7e8-kube-api-access-hhnvb\") pod \"ironic-operator-controller-manager-69d6c9f5b8-gh4kf\" (UID: \"72203b24-67d7-4dac-8430-51c940eaa7e8\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240470 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvbhc\" (UniqueName: \"kubernetes.io/projected/e5ac1213-b336-49f2-a1c3-93fbbbdc724c-kube-api-access-hvbhc\") pod \"cinder-operator-controller-manager-69cf5d4557-dlkpl\" (UID: \"e5ac1213-b336-49f2-a1c3-93fbbbdc724c\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.240516 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fj6z\" (UniqueName: \"kubernetes.io/projected/ab605e98-a012-403e-9369-aa87d02895ba-kube-api-access-7fj6z\") pod \"mariadb-operator-controller-manager-c87fff755-4547l\" (UID: \"ab605e98-a012-403e-9369-aa87d02895ba\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.243393 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.251852 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:14.75179467 +0000 UTC m=+1218.111563187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.274877 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.279847 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.280778 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.291584 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5pds\" (UniqueName: \"kubernetes.io/projected/2d3d18eb-0f1a-4552-91d3-6c6bec689f12-kube-api-access-k5pds\") pod \"heat-operator-controller-manager-594c8c9d5d-glr2h\" (UID: \"2d3d18eb-0f1a-4552-91d3-6c6bec689f12\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.292755 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-hwtp2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.294967 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvbhc\" (UniqueName: \"kubernetes.io/projected/e5ac1213-b336-49f2-a1c3-93fbbbdc724c-kube-api-access-hvbhc\") pod \"cinder-operator-controller-manager-69cf5d4557-dlkpl\" (UID: \"e5ac1213-b336-49f2-a1c3-93fbbbdc724c\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.295370 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9c8n\" (UniqueName: \"kubernetes.io/projected/62432f9a-3dc6-4f17-844e-fd5a6e983097-kube-api-access-z9c8n\") pod \"horizon-operator-controller-manager-77d5c5b54f-hfd9z\" (UID: \"62432f9a-3dc6-4f17-844e-fd5a6e983097\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.298613 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhnvb\" (UniqueName: \"kubernetes.io/projected/72203b24-67d7-4dac-8430-51c940eaa7e8-kube-api-access-hhnvb\") pod \"ironic-operator-controller-manager-69d6c9f5b8-gh4kf\" (UID: \"72203b24-67d7-4dac-8430-51c940eaa7e8\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.299347 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qggxk\" (UniqueName: \"kubernetes.io/projected/4f4c9bb1-0e84-4d29-9da3-c068c13723db-kube-api-access-qggxk\") pod \"designate-operator-controller-manager-b45d7bf98-hpfrj\" (UID: \"4f4c9bb1-0e84-4d29-9da3-c068c13723db\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.299451 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgb8j\" (UniqueName: \"kubernetes.io/projected/45ead174-1e73-410e-a2af-5db3a287e6ef-kube-api-access-pgb8j\") pod \"glance-operator-controller-manager-78fdd796fd-zxzq5\" (UID: \"45ead174-1e73-410e-a2af-5db3a287e6ef\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.299776 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7x85\" (UniqueName: \"kubernetes.io/projected/43c5f5ad-e7e7-460d-9ed2-555cc840253f-kube-api-access-b7x85\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.302932 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt4z2\" (UniqueName: \"kubernetes.io/projected/84ad66e0-8ccb-4bd1-a5c6-ec82697798ed-kube-api-access-bt4z2\") pod \"keystone-operator-controller-manager-b8b6d4659-vf8kq\" (UID: \"84ad66e0-8ccb-4bd1-a5c6-ec82697798ed\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.311025 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.309856 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.313779 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxw59\" (UniqueName: \"kubernetes.io/projected/4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf-kube-api-access-bxw59\") pod \"manila-operator-controller-manager-78c6999f6f-h5xn2\" (UID: \"4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.329237 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.332656 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.333565 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.333940 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.335259 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-c88gl" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.342197 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q44t5\" (UniqueName: \"kubernetes.io/projected/f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7-kube-api-access-q44t5\") pod \"neutron-operator-controller-manager-5d8f59fb49-9v24b\" (UID: \"f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.342315 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fj6z\" (UniqueName: \"kubernetes.io/projected/ab605e98-a012-403e-9369-aa87d02895ba-kube-api-access-7fj6z\") pod \"mariadb-operator-controller-manager-c87fff755-4547l\" (UID: \"ab605e98-a012-403e-9369-aa87d02895ba\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.342345 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf8bv\" (UniqueName: \"kubernetes.io/projected/a89f20bd-9744-47b0-abcb-07b75b44d255-kube-api-access-qf8bv\") pod \"nova-operator-controller-manager-6b8bc8d87d-pgqb5\" (UID: \"a89f20bd-9744-47b0-abcb-07b75b44d255\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.354885 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.355203 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.369480 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fj6z\" (UniqueName: \"kubernetes.io/projected/ab605e98-a012-403e-9369-aa87d02895ba-kube-api-access-7fj6z\") pod \"mariadb-operator-controller-manager-c87fff755-4547l\" (UID: \"ab605e98-a012-403e-9369-aa87d02895ba\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.369650 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.373554 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.377887 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-n7zxr" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.378080 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q44t5\" (UniqueName: \"kubernetes.io/projected/f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7-kube-api-access-q44t5\") pod \"neutron-operator-controller-manager-5d8f59fb49-9v24b\" (UID: \"f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.386570 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.388544 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.393849 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.394853 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-dc4t9" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.398768 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.408214 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.412455 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.413509 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.414899 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.419212 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.419992 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.422225 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-8hv4h" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.431139 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.434824 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.435623 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.437810 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4s984" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.440912 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447488 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c89cg\" (UniqueName: \"kubernetes.io/projected/6ede4b10-e9d1-44fe-a400-4bb900e2dc59-kube-api-access-c89cg\") pod \"swift-operator-controller-manager-547cbdb99f-s4pbs\" (UID: \"6ede4b10-e9d1-44fe-a400-4bb900e2dc59\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447538 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn7jw\" (UniqueName: \"kubernetes.io/projected/d9562ef6-7f4b-49b3-bd82-eba3360ab2b3-kube-api-access-wn7jw\") pod \"placement-operator-controller-manager-5d646b7d76-rtzb5\" (UID: \"d9562ef6-7f4b-49b3-bd82-eba3360ab2b3\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447565 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf8bv\" (UniqueName: \"kubernetes.io/projected/a89f20bd-9744-47b0-abcb-07b75b44d255-kube-api-access-qf8bv\") pod \"nova-operator-controller-manager-6b8bc8d87d-pgqb5\" (UID: \"a89f20bd-9744-47b0-abcb-07b75b44d255\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447704 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcs4c\" (UniqueName: \"kubernetes.io/projected/0fbc160a-787a-4a04-ab60-ddfa07e6d6ca-kube-api-access-jcs4c\") pod \"ovn-operator-controller-manager-55db956ddc-2dctq\" (UID: \"0fbc160a-787a-4a04-ab60-ddfa07e6d6ca\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447861 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tk4j\" (UniqueName: \"kubernetes.io/projected/2e57fccc-cded-48d6-bd9a-c797d2013d6d-kube-api-access-5tk4j\") pod \"octavia-operator-controller-manager-7bd9774b6-vxfhs\" (UID: \"2e57fccc-cded-48d6-bd9a-c797d2013d6d\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.447945 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.448036 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbdhl\" (UniqueName: \"kubernetes.io/projected/19f29c0b-c9b6-4137-9558-6af7128aa0b9-kube-api-access-dbdhl\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.464126 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.466226 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.468252 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.482354 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.490777 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.492227 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-286hb" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.493585 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.494342 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.498094 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-ts5cr" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.498151 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf8bv\" (UniqueName: \"kubernetes.io/projected/a89f20bd-9744-47b0-abcb-07b75b44d255-kube-api-access-qf8bv\") pod \"nova-operator-controller-manager-6b8bc8d87d-pgqb5\" (UID: \"a89f20bd-9744-47b0-abcb-07b75b44d255\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.503973 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.518609 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.542607 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.557992 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c89cg\" (UniqueName: \"kubernetes.io/projected/6ede4b10-e9d1-44fe-a400-4bb900e2dc59-kube-api-access-c89cg\") pod \"swift-operator-controller-manager-547cbdb99f-s4pbs\" (UID: \"6ede4b10-e9d1-44fe-a400-4bb900e2dc59\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558072 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2ghh\" (UniqueName: \"kubernetes.io/projected/34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed-kube-api-access-k2ghh\") pod \"telemetry-operator-controller-manager-85cd9769bb-8h7wm\" (UID: \"34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558100 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn7jw\" (UniqueName: \"kubernetes.io/projected/d9562ef6-7f4b-49b3-bd82-eba3360ab2b3-kube-api-access-wn7jw\") pod \"placement-operator-controller-manager-5d646b7d76-rtzb5\" (UID: \"d9562ef6-7f4b-49b3-bd82-eba3360ab2b3\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558153 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcs4c\" (UniqueName: \"kubernetes.io/projected/0fbc160a-787a-4a04-ab60-ddfa07e6d6ca-kube-api-access-jcs4c\") pod \"ovn-operator-controller-manager-55db956ddc-2dctq\" (UID: \"0fbc160a-787a-4a04-ab60-ddfa07e6d6ca\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558212 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tk4j\" (UniqueName: \"kubernetes.io/projected/2e57fccc-cded-48d6-bd9a-c797d2013d6d-kube-api-access-5tk4j\") pod \"octavia-operator-controller-manager-7bd9774b6-vxfhs\" (UID: \"2e57fccc-cded-48d6-bd9a-c797d2013d6d\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558238 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bbvg\" (UniqueName: \"kubernetes.io/projected/ada6a395-b944-4798-bb31-347fbc476be4-kube-api-access-9bbvg\") pod \"test-operator-controller-manager-69797bbcbd-vq98j\" (UID: \"ada6a395-b944-4798-bb31-347fbc476be4\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558263 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.558308 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbdhl\" (UniqueName: \"kubernetes.io/projected/19f29c0b-c9b6-4137-9558-6af7128aa0b9-kube-api-access-dbdhl\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.558862 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.558900 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:15.058887866 +0000 UTC m=+1218.418656383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.560803 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.561862 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.568295 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.577440 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-hjcpk" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.586371 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn7jw\" (UniqueName: \"kubernetes.io/projected/d9562ef6-7f4b-49b3-bd82-eba3360ab2b3-kube-api-access-wn7jw\") pod \"placement-operator-controller-manager-5d646b7d76-rtzb5\" (UID: \"d9562ef6-7f4b-49b3-bd82-eba3360ab2b3\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.592995 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcs4c\" (UniqueName: \"kubernetes.io/projected/0fbc160a-787a-4a04-ab60-ddfa07e6d6ca-kube-api-access-jcs4c\") pod \"ovn-operator-controller-manager-55db956ddc-2dctq\" (UID: \"0fbc160a-787a-4a04-ab60-ddfa07e6d6ca\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.598314 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbdhl\" (UniqueName: \"kubernetes.io/projected/19f29c0b-c9b6-4137-9558-6af7128aa0b9-kube-api-access-dbdhl\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.600407 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tk4j\" (UniqueName: \"kubernetes.io/projected/2e57fccc-cded-48d6-bd9a-c797d2013d6d-kube-api-access-5tk4j\") pod \"octavia-operator-controller-manager-7bd9774b6-vxfhs\" (UID: \"2e57fccc-cded-48d6-bd9a-c797d2013d6d\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.617505 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c89cg\" (UniqueName: \"kubernetes.io/projected/6ede4b10-e9d1-44fe-a400-4bb900e2dc59-kube-api-access-c89cg\") pod \"swift-operator-controller-manager-547cbdb99f-s4pbs\" (UID: \"6ede4b10-e9d1-44fe-a400-4bb900e2dc59\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.622576 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.646984 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.648185 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.667839 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2ghh\" (UniqueName: \"kubernetes.io/projected/34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed-kube-api-access-k2ghh\") pod \"telemetry-operator-controller-manager-85cd9769bb-8h7wm\" (UID: \"34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.667979 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wng6n\" (UniqueName: \"kubernetes.io/projected/155c2cd6-e900-4e96-bc0b-2b3227732430-kube-api-access-wng6n\") pod \"watcher-operator-controller-manager-5ffb9c6597-628hc\" (UID: \"155c2cd6-e900-4e96-bc0b-2b3227732430\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.668029 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bbvg\" (UniqueName: \"kubernetes.io/projected/ada6a395-b944-4798-bb31-347fbc476be4-kube-api-access-9bbvg\") pod \"test-operator-controller-manager-69797bbcbd-vq98j\" (UID: \"ada6a395-b944-4798-bb31-347fbc476be4\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.672753 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.674308 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.674545 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-vqm7d" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.709056 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.724686 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2ghh\" (UniqueName: \"kubernetes.io/projected/34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed-kube-api-access-k2ghh\") pod \"telemetry-operator-controller-manager-85cd9769bb-8h7wm\" (UID: \"34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.739523 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.741715 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bbvg\" (UniqueName: \"kubernetes.io/projected/ada6a395-b944-4798-bb31-347fbc476be4-kube-api-access-9bbvg\") pod \"test-operator-controller-manager-69797bbcbd-vq98j\" (UID: \"ada6a395-b944-4798-bb31-347fbc476be4\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.756291 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.760259 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.764592 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.765669 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-4l4tk" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.771635 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r"] Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.771651 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.771886 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.771978 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:15.771962618 +0000 UTC m=+1219.131731135 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.772666 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.772750 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wng6n\" (UniqueName: \"kubernetes.io/projected/155c2cd6-e900-4e96-bc0b-2b3227732430-kube-api-access-wng6n\") pod \"watcher-operator-controller-manager-5ffb9c6597-628hc\" (UID: \"155c2cd6-e900-4e96-bc0b-2b3227732430\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.772793 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5b7l\" (UniqueName: \"kubernetes.io/projected/fddde490-1908-48b2-afbc-d09bb4ec8e4a-kube-api-access-w5b7l\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.772862 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.797593 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wng6n\" (UniqueName: \"kubernetes.io/projected/155c2cd6-e900-4e96-bc0b-2b3227732430-kube-api-access-wng6n\") pod \"watcher-operator-controller-manager-5ffb9c6597-628hc\" (UID: \"155c2cd6-e900-4e96-bc0b-2b3227732430\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.805642 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.853449 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.876998 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.877233 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.877274 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5b7l\" (UniqueName: \"kubernetes.io/projected/fddde490-1908-48b2-afbc-d09bb4ec8e4a-kube-api-access-w5b7l\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.877413 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:15.377272161 +0000 UTC m=+1218.737040678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.877505 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.877630 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hndwb\" (UniqueName: \"kubernetes.io/projected/8a4951f8-3310-4e38-b989-fa900a0e68bd-kube-api-access-hndwb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gkb6r\" (UID: \"8a4951f8-3310-4e38-b989-fa900a0e68bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.877690 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.877700 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:02:14 crc kubenswrapper[4836]: E0122 10:02:14.877726 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:15.377716083 +0000 UTC m=+1218.737484600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.896620 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.939780 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5b7l\" (UniqueName: \"kubernetes.io/projected/fddde490-1908-48b2-afbc-d09bb4ec8e4a-kube-api-access-w5b7l\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.978465 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:02:14 crc kubenswrapper[4836]: I0122 10:02:14.978939 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hndwb\" (UniqueName: \"kubernetes.io/projected/8a4951f8-3310-4e38-b989-fa900a0e68bd-kube-api-access-hndwb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gkb6r\" (UID: \"8a4951f8-3310-4e38-b989-fa900a0e68bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.018243 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hndwb\" (UniqueName: \"kubernetes.io/projected/8a4951f8-3310-4e38-b989-fa900a0e68bd-kube-api-access-hndwb\") pod \"rabbitmq-cluster-operator-manager-668c99d594-gkb6r\" (UID: \"8a4951f8-3310-4e38-b989-fa900a0e68bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.081711 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.082678 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.082719 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:16.082704596 +0000 UTC m=+1219.442473113 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.266348 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.334626 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl"] Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.354568 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj"] Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.373900 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" event={"ID":"e5ac1213-b336-49f2-a1c3-93fbbbdc724c","Type":"ContainerStarted","Data":"45584afee3f9571e2c2e0647ba55c538d4362ddb9aae65f66ef4827a4b226cb8"} Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.386805 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.386915 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.387077 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.387125 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:16.387108423 +0000 UTC m=+1219.746876940 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.387430 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.387459 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:16.387451862 +0000 UTC m=+1219.747220379 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.421463 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq"] Jan 22 10:02:15 crc kubenswrapper[4836]: I0122 10:02:15.800000 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.800482 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:15 crc kubenswrapper[4836]: E0122 10:02:15.800546 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:17.800529102 +0000 UTC m=+1221.160297619 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.097672 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.105666 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.106125 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.106173 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:18.106158482 +0000 UTC m=+1221.465926999 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.108591 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45ead174_1e73_410e_a2af_5db3a287e6ef.slice/crio-2cd395359e1ac77d59c62785b52cc52b8efd8864986cee8d492c3e3a3d8233e2 WatchSource:0}: Error finding container 2cd395359e1ac77d59c62785b52cc52b8efd8864986cee8d492c3e3a3d8233e2: Status 404 returned error can't find the container with id 2cd395359e1ac77d59c62785b52cc52b8efd8864986cee8d492c3e3a3d8233e2 Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.129675 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.160208 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h"] Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.164103 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fbc160a_787a_4a04_ab60_ddfa07e6d6ca.slice/crio-43a3e7a54c552a1b4fd14d583eaf9b2a4aad0cbdbf5494cd93d0ab15362ab0f7 WatchSource:0}: Error finding container 43a3e7a54c552a1b4fd14d583eaf9b2a4aad0cbdbf5494cd93d0ab15362ab0f7: Status 404 returned error can't find the container with id 43a3e7a54c552a1b4fd14d583eaf9b2a4aad0cbdbf5494cd93d0ab15362ab0f7 Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.172987 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.186127 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.197172 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.204137 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.224836 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.234245 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq"] Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.236378 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda89f20bd_9744_47b0_abcb_07b75b44d255.slice/crio-e9daf887d91be0cb4e48dda99eec2afd6fb8e602b4af7df4c1db8ce89dc79406 WatchSource:0}: Error finding container e9daf887d91be0cb4e48dda99eec2afd6fb8e602b4af7df4c1db8ce89dc79406: Status 404 returned error can't find the container with id e9daf887d91be0cb4e48dda99eec2afd6fb8e602b4af7df4c1db8ce89dc79406 Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.240346 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs"] Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.243663 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e57fccc_cded_48d6_bd9a_c797d2013d6d.slice/crio-f1db428e375592d078dbb0a9fdc2d305d5d288b422b30a0e5f0b094ea98f62c1 WatchSource:0}: Error finding container f1db428e375592d078dbb0a9fdc2d305d5d288b422b30a0e5f0b094ea98f62c1: Status 404 returned error can't find the container with id f1db428e375592d078dbb0a9fdc2d305d5d288b422b30a0e5f0b094ea98f62c1 Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.418600 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.418723 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.418855 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.418877 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.418904 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:18.418888934 +0000 UTC m=+1221.778657451 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.418950 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:18.418930265 +0000 UTC m=+1221.778698842 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.433245 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" event={"ID":"2d3d18eb-0f1a-4552-91d3-6c6bec689f12","Type":"ContainerStarted","Data":"f5069e540191945d47c5b1a24aae1cc390f6654824653d3abbbca7e720191880"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.435073 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" event={"ID":"2e57fccc-cded-48d6-bd9a-c797d2013d6d","Type":"ContainerStarted","Data":"f1db428e375592d078dbb0a9fdc2d305d5d288b422b30a0e5f0b094ea98f62c1"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.436694 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" event={"ID":"a89f20bd-9744-47b0-abcb-07b75b44d255","Type":"ContainerStarted","Data":"e9daf887d91be0cb4e48dda99eec2afd6fb8e602b4af7df4c1db8ce89dc79406"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.438205 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" event={"ID":"0fbc160a-787a-4a04-ab60-ddfa07e6d6ca","Type":"ContainerStarted","Data":"43a3e7a54c552a1b4fd14d583eaf9b2a4aad0cbdbf5494cd93d0ab15362ab0f7"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.448466 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" event={"ID":"f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7","Type":"ContainerStarted","Data":"a6c95d82a7f35e28623f46fe8a9bca1d5c18f51d84df8e87d4f5e7ff94b7da89"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.455409 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" event={"ID":"ab605e98-a012-403e-9369-aa87d02895ba","Type":"ContainerStarted","Data":"d44ae31f980b4a2397f4ad53baa10b39053c4e110588bd66902aa011b82d444e"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.457883 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" event={"ID":"bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b","Type":"ContainerStarted","Data":"1aeefc69f99f207430e05d4dc372caae6e362dd1b48ddb3cf4a023fa77540cca"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.459448 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" event={"ID":"84ad66e0-8ccb-4bd1-a5c6-ec82697798ed","Type":"ContainerStarted","Data":"82d4fa2f4eb5a1147c1f72d17591be9b175bf075a54b0f9870231bef7b8dc933"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.460435 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" event={"ID":"45ead174-1e73-410e-a2af-5db3a287e6ef","Type":"ContainerStarted","Data":"2cd395359e1ac77d59c62785b52cc52b8efd8864986cee8d492c3e3a3d8233e2"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.461343 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" event={"ID":"4f4c9bb1-0e84-4d29-9da3-c068c13723db","Type":"ContainerStarted","Data":"3a2b8dee9343e4b93d9aa430ec15c8f174c0a7902c0d945defee02263098ecb9"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.462084 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" event={"ID":"6ede4b10-e9d1-44fe-a400-4bb900e2dc59","Type":"ContainerStarted","Data":"aeec693eff9d88601ea732be361ca038b26bac50fe1c407f742550d203728101"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.463639 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" event={"ID":"72203b24-67d7-4dac-8430-51c940eaa7e8","Type":"ContainerStarted","Data":"bc481afcbeca9d8208931dfb714f6e337c3881723832ee107daa381ba08b45dc"} Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.479712 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.489886 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.497092 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc"] Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.506019 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j"] Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.506722 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34b2d45c_eab1_4f20_9ddb_fb6769fdd4ed.slice/crio-b14fd01ae1ab2f9c26510e1de1a93eb6d47265e857b2355b20c0d41423918555 WatchSource:0}: Error finding container b14fd01ae1ab2f9c26510e1de1a93eb6d47265e857b2355b20c0d41423918555: Status 404 returned error can't find the container with id b14fd01ae1ab2f9c26510e1de1a93eb6d47265e857b2355b20c0d41423918555 Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.508674 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podada6a395_b944_4798_bb31_347fbc476be4.slice/crio-b4f173cc835c525e4b5465575e295a5bb1c652b3d73f86b334373ca700971fa5 WatchSource:0}: Error finding container b4f173cc835c525e4b5465575e295a5bb1c652b3d73f86b334373ca700971fa5: Status 404 returned error can't find the container with id b4f173cc835c525e4b5465575e295a5bb1c652b3d73f86b334373ca700971fa5 Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.509606 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k2ghh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-85cd9769bb-8h7wm_openstack-operators(34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.510721 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" podUID="34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.510831 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9bbvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-vq98j_openstack-operators(ada6a395-b944-4798-bb31-347fbc476be4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.511739 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod155c2cd6_e900_4e96_bc0b_2b3227732430.slice/crio-b262fc8f4465063f979ab31b6935a41f5b597f6f357d31d5dc4a8a591349c40c WatchSource:0}: Error finding container b262fc8f4465063f979ab31b6935a41f5b597f6f357d31d5dc4a8a591349c40c: Status 404 returned error can't find the container with id b262fc8f4465063f979ab31b6935a41f5b597f6f357d31d5dc4a8a591349c40c Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.511975 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cdbaa6a_2b38_4350_87c9_e45bf3cc89cf.slice/crio-d12aad06ae825bd66098ba4fb55ce0e8ce6eaec1699c9c714ad6675318477538 WatchSource:0}: Error finding container d12aad06ae825bd66098ba4fb55ce0e8ce6eaec1699c9c714ad6675318477538: Status 404 returned error can't find the container with id d12aad06ae825bd66098ba4fb55ce0e8ce6eaec1699c9c714ad6675318477538 Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.512564 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2"] Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.513835 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" podUID="ada6a395-b944-4798-bb31-347fbc476be4" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.515532 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wng6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5ffb9c6597-628hc_openstack-operators(155c2cd6-e900-4e96-bc0b-2b3227732430): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.516026 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxw59,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-78c6999f6f-h5xn2_openstack-operators(4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.516682 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" podUID="155c2cd6-e900-4e96-bc0b-2b3227732430" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.517476 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" podUID="4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf" Jan 22 10:02:16 crc kubenswrapper[4836]: W0122 10:02:16.517612 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a4951f8_3310_4e38_b989_fa900a0e68bd.slice/crio-fec0fecea2b62a65bf3138a9f6aede2dfdb9b8dfc084ce6e6287d9bb8d474b6c WatchSource:0}: Error finding container fec0fecea2b62a65bf3138a9f6aede2dfdb9b8dfc084ce6e6287d9bb8d474b6c: Status 404 returned error can't find the container with id fec0fecea2b62a65bf3138a9f6aede2dfdb9b8dfc084ce6e6287d9bb8d474b6c Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.517675 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z9c8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-77d5c5b54f-hfd9z_openstack-operators(62432f9a-3dc6-4f17-844e-fd5a6e983097): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.518908 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" podUID="62432f9a-3dc6-4f17-844e-fd5a6e983097" Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.518957 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z"] Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.521276 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hndwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-gkb6r_openstack-operators(8a4951f8-3310-4e38-b989-fa900a0e68bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 22 10:02:16 crc kubenswrapper[4836]: I0122 10:02:16.524102 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r"] Jan 22 10:02:16 crc kubenswrapper[4836]: E0122 10:02:16.524193 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" podUID="8a4951f8-3310-4e38-b989-fa900a0e68bd" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.474328 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" event={"ID":"155c2cd6-e900-4e96-bc0b-2b3227732430","Type":"ContainerStarted","Data":"b262fc8f4465063f979ab31b6935a41f5b597f6f357d31d5dc4a8a591349c40c"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.477287 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" podUID="155c2cd6-e900-4e96-bc0b-2b3227732430" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.485732 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" event={"ID":"8a4951f8-3310-4e38-b989-fa900a0e68bd","Type":"ContainerStarted","Data":"fec0fecea2b62a65bf3138a9f6aede2dfdb9b8dfc084ce6e6287d9bb8d474b6c"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.490246 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" podUID="8a4951f8-3310-4e38-b989-fa900a0e68bd" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.495087 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" event={"ID":"34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed","Type":"ContainerStarted","Data":"b14fd01ae1ab2f9c26510e1de1a93eb6d47265e857b2355b20c0d41423918555"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.496430 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" podUID="34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.499629 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" event={"ID":"4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf","Type":"ContainerStarted","Data":"d12aad06ae825bd66098ba4fb55ce0e8ce6eaec1699c9c714ad6675318477538"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.506546 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" podUID="4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.510737 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" event={"ID":"ada6a395-b944-4798-bb31-347fbc476be4","Type":"ContainerStarted","Data":"b4f173cc835c525e4b5465575e295a5bb1c652b3d73f86b334373ca700971fa5"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.512822 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" podUID="ada6a395-b944-4798-bb31-347fbc476be4" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.514249 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" event={"ID":"d9562ef6-7f4b-49b3-bd82-eba3360ab2b3","Type":"ContainerStarted","Data":"b57020d6f660279113ee94518944894ca2bf83f4099ff72584ceb24b11d31c6b"} Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.517305 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" event={"ID":"62432f9a-3dc6-4f17-844e-fd5a6e983097","Type":"ContainerStarted","Data":"103d7f72a25c870a2f4b612e58c19bb28d4caa67a9699187d0beaabb6e2e71da"} Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.518773 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" podUID="62432f9a-3dc6-4f17-844e-fd5a6e983097" Jan 22 10:02:17 crc kubenswrapper[4836]: I0122 10:02:17.844700 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.844846 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:17 crc kubenswrapper[4836]: E0122 10:02:17.844925 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:21.844906785 +0000 UTC m=+1225.204675302 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: I0122 10:02:18.160989 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.161168 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.161250 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:22.161231265 +0000 UTC m=+1225.520999782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: I0122 10:02:18.464504 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:18 crc kubenswrapper[4836]: I0122 10:02:18.464586 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.464740 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.464794 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:22.464777037 +0000 UTC m=+1225.824545554 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.464858 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.464883 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:22.46487518 +0000 UTC m=+1225.824643697 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.534874 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" podUID="34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.534945 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" podUID="155c2cd6-e900-4e96-bc0b-2b3227732430" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.534991 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" podUID="ada6a395-b944-4798-bb31-347fbc476be4" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.535078 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" podUID="62432f9a-3dc6-4f17-844e-fd5a6e983097" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.535329 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" podUID="4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf" Jan 22 10:02:18 crc kubenswrapper[4836]: E0122 10:02:18.535674 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" podUID="8a4951f8-3310-4e38-b989-fa900a0e68bd" Jan 22 10:02:21 crc kubenswrapper[4836]: I0122 10:02:21.926868 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:21 crc kubenswrapper[4836]: E0122 10:02:21.927119 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:21 crc kubenswrapper[4836]: E0122 10:02:21.927335 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:29.927315249 +0000 UTC m=+1233.287083756 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: I0122 10:02:22.229876 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.230086 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.230153 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:30.230134312 +0000 UTC m=+1233.589902829 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: I0122 10:02:22.535255 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:22 crc kubenswrapper[4836]: I0122 10:02:22.535713 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.535428 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.535788 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:30.535769963 +0000 UTC m=+1233.895538480 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.535960 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:22 crc kubenswrapper[4836]: E0122 10:02:22.536009 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:30.535994629 +0000 UTC m=+1233.895763146 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:23 crc kubenswrapper[4836]: I0122 10:02:23.897468 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:02:23 crc kubenswrapper[4836]: I0122 10:02:23.897547 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:02:28 crc kubenswrapper[4836]: I0122 10:02:28.999321 4836 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 10:02:29 crc kubenswrapper[4836]: I0122 10:02:29.985341 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:29 crc kubenswrapper[4836]: E0122 10:02:29.985671 4836 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:29 crc kubenswrapper[4836]: E0122 10:02:29.985740 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert podName:43c5f5ad-e7e7-460d-9ed2-555cc840253f nodeName:}" failed. No retries permitted until 2026-01-22 10:02:45.985720444 +0000 UTC m=+1249.345488971 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert") pod "infra-operator-controller-manager-54ccf4f85d-4p2bd" (UID: "43c5f5ad-e7e7-460d-9ed2-555cc840253f") : secret "infra-operator-webhook-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: I0122 10:02:30.290076 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.290250 4836 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.290334 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert podName:19f29c0b-c9b6-4137-9558-6af7128aa0b9 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:46.290316425 +0000 UTC m=+1249.650084942 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" (UID: "19f29c0b-c9b6-4137-9558-6af7128aa0b9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: I0122 10:02:30.593402 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.593551 4836 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.593724 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:46.593703796 +0000 UTC m=+1249.953472313 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "webhook-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: I0122 10:02:30.594225 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.594313 4836 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 22 10:02:30 crc kubenswrapper[4836]: E0122 10:02:30.594338 4836 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs podName:fddde490-1908-48b2-afbc-d09bb4ec8e4a nodeName:}" failed. No retries permitted until 2026-01-22 10:02:46.594330593 +0000 UTC m=+1249.954099110 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-jtvkz" (UID: "fddde490-1908-48b2-afbc-d09bb4ec8e4a") : secret "metrics-server-cert" not found Jan 22 10:02:31 crc kubenswrapper[4836]: E0122 10:02:31.345179 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71" Jan 22 10:02:31 crc kubenswrapper[4836]: E0122 10:02:31.345621 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7fj6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-c87fff755-4547l_openstack-operators(ab605e98-a012-403e-9369-aa87d02895ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:31 crc kubenswrapper[4836]: E0122 10:02:31.346865 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" podUID="ab605e98-a012-403e-9369-aa87d02895ba" Jan 22 10:02:31 crc kubenswrapper[4836]: E0122 10:02:31.633306 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" podUID="ab605e98-a012-403e-9369-aa87d02895ba" Jan 22 10:02:32 crc kubenswrapper[4836]: E0122 10:02:32.553555 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece" Jan 22 10:02:32 crc kubenswrapper[4836]: E0122 10:02:32.553784 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qggxk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-b45d7bf98-hpfrj_openstack-operators(4f4c9bb1-0e84-4d29-9da3-c068c13723db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:32 crc kubenswrapper[4836]: E0122 10:02:32.554986 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" podUID="4f4c9bb1-0e84-4d29-9da3-c068c13723db" Jan 22 10:02:32 crc kubenswrapper[4836]: E0122 10:02:32.639472 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece\\\"\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" podUID="4f4c9bb1-0e84-4d29-9da3-c068c13723db" Jan 22 10:02:33 crc kubenswrapper[4836]: E0122 10:02:33.463735 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0" Jan 22 10:02:33 crc kubenswrapper[4836]: E0122 10:02:33.463991 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wn7jw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5d646b7d76-rtzb5_openstack-operators(d9562ef6-7f4b-49b3-bd82-eba3360ab2b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:33 crc kubenswrapper[4836]: E0122 10:02:33.465205 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" podUID="d9562ef6-7f4b-49b3-bd82-eba3360ab2b3" Jan 22 10:02:33 crc kubenswrapper[4836]: E0122 10:02:33.646836 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" podUID="d9562ef6-7f4b-49b3-bd82-eba3360ab2b3" Jan 22 10:02:34 crc kubenswrapper[4836]: E0122 10:02:34.191335 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5" Jan 22 10:02:34 crc kubenswrapper[4836]: E0122 10:02:34.191567 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5tk4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7bd9774b6-vxfhs_openstack-operators(2e57fccc-cded-48d6-bd9a-c797d2013d6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:34 crc kubenswrapper[4836]: E0122 10:02:34.192886 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" podUID="2e57fccc-cded-48d6-bd9a-c797d2013d6d" Jan 22 10:02:34 crc kubenswrapper[4836]: E0122 10:02:34.651140 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" podUID="2e57fccc-cded-48d6-bd9a-c797d2013d6d" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.075583 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.075883 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5pds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-594c8c9d5d-glr2h_openstack-operators(2d3d18eb-0f1a-4552-91d3-6c6bec689f12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.077782 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" podUID="2d3d18eb-0f1a-4552-91d3-6c6bec689f12" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.662251 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492\\\"\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" podUID="2d3d18eb-0f1a-4552-91d3-6c6bec689f12" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.959739 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:d3c55b59cb192799f8d31196c55c9e9bb3cd38aef7ec51ef257dabf1548e8b30" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.959977 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:d3c55b59cb192799f8d31196c55c9e9bb3cd38aef7ec51ef257dabf1548e8b30,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hhnvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-69d6c9f5b8-gh4kf_openstack-operators(72203b24-67d7-4dac-8430-51c940eaa7e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:35 crc kubenswrapper[4836]: E0122 10:02:35.961210 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" podUID="72203b24-67d7-4dac-8430-51c940eaa7e8" Jan 22 10:02:36 crc kubenswrapper[4836]: E0122 10:02:36.673938 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:d3c55b59cb192799f8d31196c55c9e9bb3cd38aef7ec51ef257dabf1548e8b30\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" podUID="72203b24-67d7-4dac-8430-51c940eaa7e8" Jan 22 10:02:37 crc kubenswrapper[4836]: E0122 10:02:37.003083 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4" Jan 22 10:02:37 crc kubenswrapper[4836]: E0122 10:02:37.003587 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q44t5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5d8f59fb49-9v24b_openstack-operators(f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:37 crc kubenswrapper[4836]: E0122 10:02:37.005055 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" podUID="f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7" Jan 22 10:02:37 crc kubenswrapper[4836]: E0122 10:02:37.678697 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" podUID="f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.269251 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.269997 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c89cg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-547cbdb99f-s4pbs_openstack-operators(6ede4b10-e9d1-44fe-a400-4bb900e2dc59): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.272026 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" podUID="6ede4b10-e9d1-44fe-a400-4bb900e2dc59" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.702529 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" podUID="6ede4b10-e9d1-44fe-a400-4bb900e2dc59" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.904225 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.904527 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p2855,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-59dd8b7cbf-7dnbq_openstack-operators(bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:41 crc kubenswrapper[4836]: E0122 10:02:41.905714 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" podUID="bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b" Jan 22 10:02:42 crc kubenswrapper[4836]: E0122 10:02:42.707968 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" podUID="bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b" Jan 22 10:02:44 crc kubenswrapper[4836]: E0122 10:02:44.817440 4836 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349" Jan 22 10:02:44 crc kubenswrapper[4836]: E0122 10:02:44.817637 4836 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bt4z2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b8b6d4659-vf8kq_openstack-operators(84ad66e0-8ccb-4bd1-a5c6-ec82697798ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 22 10:02:44 crc kubenswrapper[4836]: E0122 10:02:44.818831 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" podUID="84ad66e0-8ccb-4bd1-a5c6-ec82697798ed" Jan 22 10:02:45 crc kubenswrapper[4836]: E0122 10:02:45.728846 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" podUID="84ad66e0-8ccb-4bd1-a5c6-ec82697798ed" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.043638 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.054427 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/43c5f5ad-e7e7-460d-9ed2-555cc840253f-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-4p2bd\" (UID: \"43c5f5ad-e7e7-460d-9ed2-555cc840253f\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.244920 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.349014 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.353905 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19f29c0b-c9b6-4137-9558-6af7128aa0b9-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs\" (UID: \"19f29c0b-c9b6-4137-9558-6af7128aa0b9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.632995 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.653360 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.653553 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.659267 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.661778 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/fddde490-1908-48b2-afbc-d09bb4ec8e4a-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-jtvkz\" (UID: \"fddde490-1908-48b2-afbc-d09bb4ec8e4a\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:46 crc kubenswrapper[4836]: I0122 10:02:46.736728 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:02:53 crc kubenswrapper[4836]: I0122 10:02:53.897108 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:02:53 crc kubenswrapper[4836]: I0122 10:02:53.897471 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:02:53 crc kubenswrapper[4836]: I0122 10:02:53.897616 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 10:02:53 crc kubenswrapper[4836]: I0122 10:02:53.898336 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 10:02:53 crc kubenswrapper[4836]: I0122 10:02:53.898405 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067" gracePeriod=600 Jan 22 10:03:09 crc kubenswrapper[4836]: I0122 10:03:09.890804 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067" exitCode=0 Jan 22 10:03:09 crc kubenswrapper[4836]: I0122 10:03:09.891012 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067"} Jan 22 10:03:09 crc kubenswrapper[4836]: I0122 10:03:09.891618 4836 scope.go:117] "RemoveContainer" containerID="c24bbc2af97b98756aa3ea711155e36e050c48ebceaadd826f4505ebf15223c7" Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.825474 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd"] Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.921529 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62"} Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.932930 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz"] Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.935302 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" event={"ID":"0fbc160a-787a-4a04-ab60-ddfa07e6d6ca","Type":"ContainerStarted","Data":"b7ad88dfb1b55ff9d76cbbfc63d90edbea8a6c8d7e41fc04bffb2ee4250f5223"} Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.937279 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.956600 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs"] Jan 22 10:03:13 crc kubenswrapper[4836]: I0122 10:03:13.957754 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" podStartSLOduration=32.122895613 podStartE2EDuration="59.957736992s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.17694015 +0000 UTC m=+1219.536708667" lastFinishedPulling="2026-01-22 10:02:44.011781529 +0000 UTC m=+1247.371550046" observedRunningTime="2026-01-22 10:03:13.954661019 +0000 UTC m=+1277.314429536" watchObservedRunningTime="2026-01-22 10:03:13.957736992 +0000 UTC m=+1277.317505509" Jan 22 10:03:14 crc kubenswrapper[4836]: W0122 10:03:14.062180 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43c5f5ad_e7e7_460d_9ed2_555cc840253f.slice/crio-da803bdced3c96f8b745b08fd277afe9f3300c14fa55d7e16c4eb20f97741120 WatchSource:0}: Error finding container da803bdced3c96f8b745b08fd277afe9f3300c14fa55d7e16c4eb20f97741120: Status 404 returned error can't find the container with id da803bdced3c96f8b745b08fd277afe9f3300c14fa55d7e16c4eb20f97741120 Jan 22 10:03:14 crc kubenswrapper[4836]: W0122 10:03:14.065307 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfddde490_1908_48b2_afbc_d09bb4ec8e4a.slice/crio-cb9d40e3fc0c37d62786bc01102ba2a1533dd1b79ce56102c26c3365ca29c555 WatchSource:0}: Error finding container cb9d40e3fc0c37d62786bc01102ba2a1533dd1b79ce56102c26c3365ca29c555: Status 404 returned error can't find the container with id cb9d40e3fc0c37d62786bc01102ba2a1533dd1b79ce56102c26c3365ca29c555 Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.957631 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" event={"ID":"e5ac1213-b336-49f2-a1c3-93fbbbdc724c","Type":"ContainerStarted","Data":"45e17750befcf6dfe831df22229d6ecd9119a2b316e1596ad754302c7628b0eb"} Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.959087 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.969423 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" event={"ID":"4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf","Type":"ContainerStarted","Data":"a42f2ebcedf9e7463a5c512103d5a427c817e0a23fdab0ca14e607045a7933ba"} Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.970097 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.976580 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" event={"ID":"2e57fccc-cded-48d6-bd9a-c797d2013d6d","Type":"ContainerStarted","Data":"d1d748ddfd41266cb67d47dc78b5522069f81d7cb527e922b6ad4a4a17ffcd05"} Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.977203 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.980610 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" event={"ID":"19f29c0b-c9b6-4137-9558-6af7128aa0b9","Type":"ContainerStarted","Data":"90aba8c883611a2ab05eb9d3b1c76e1ff528c2373caeb4c247ec712d7089c76f"} Jan 22 10:03:14 crc kubenswrapper[4836]: I0122 10:03:14.989455 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" podStartSLOduration=33.311996845 podStartE2EDuration="1m1.989431171s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:15.334338453 +0000 UTC m=+1218.694106970" lastFinishedPulling="2026-01-22 10:02:44.011772779 +0000 UTC m=+1247.371541296" observedRunningTime="2026-01-22 10:03:14.982408981 +0000 UTC m=+1278.342177498" watchObservedRunningTime="2026-01-22 10:03:14.989431171 +0000 UTC m=+1278.349199688" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.022276 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" event={"ID":"ab605e98-a012-403e-9369-aa87d02895ba","Type":"ContainerStarted","Data":"5c23044a2268e5bd85cd7d07215bb324e30897c0b85d450621cc6ad3305053ec"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.022985 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.031501 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" event={"ID":"2d3d18eb-0f1a-4552-91d3-6c6bec689f12","Type":"ContainerStarted","Data":"d0e452fdf9048301ba7ab98f9547a8897f3a2ac105afacffc4bb01ec4f25b09a"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.032234 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.033125 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" podStartSLOduration=3.718531116 podStartE2EDuration="1m1.033104771s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.249895086 +0000 UTC m=+1219.609663603" lastFinishedPulling="2026-01-22 10:03:13.564468741 +0000 UTC m=+1276.924237258" observedRunningTime="2026-01-22 10:03:15.023276956 +0000 UTC m=+1278.383045473" watchObservedRunningTime="2026-01-22 10:03:15.033104771 +0000 UTC m=+1278.392873288" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.038339 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" event={"ID":"ada6a395-b944-4798-bb31-347fbc476be4","Type":"ContainerStarted","Data":"d9970384c73e8603742ae8a244cb3896089e920a4489df7b618f898bf72efdda"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.039506 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.052371 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" event={"ID":"43c5f5ad-e7e7-460d-9ed2-555cc840253f","Type":"ContainerStarted","Data":"da803bdced3c96f8b745b08fd277afe9f3300c14fa55d7e16c4eb20f97741120"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.074394 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" podStartSLOduration=4.148900694 podStartE2EDuration="1m1.074370837s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.514814532 +0000 UTC m=+1219.874583049" lastFinishedPulling="2026-01-22 10:03:13.440284665 +0000 UTC m=+1276.800053192" observedRunningTime="2026-01-22 10:03:15.061601402 +0000 UTC m=+1278.421369929" watchObservedRunningTime="2026-01-22 10:03:15.074370837 +0000 UTC m=+1278.434139354" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.087172 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" event={"ID":"155c2cd6-e900-4e96-bc0b-2b3227732430","Type":"ContainerStarted","Data":"176c1dd0ee75ed662c60a26d6ae5e92673a233e43a98e8ae50a7958f13c54d61"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.087953 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.108266 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" podStartSLOduration=4.19939533 podStartE2EDuration="1m1.108242022s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.510712012 +0000 UTC m=+1219.870480529" lastFinishedPulling="2026-01-22 10:03:13.419558704 +0000 UTC m=+1276.779327221" observedRunningTime="2026-01-22 10:03:15.106345402 +0000 UTC m=+1278.466113919" watchObservedRunningTime="2026-01-22 10:03:15.108242022 +0000 UTC m=+1278.468010539" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.121170 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" event={"ID":"34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed","Type":"ContainerStarted","Data":"5260d775200b00313c5b29fea3d15f0d5884ee62cc6a78556361cb2a0330edc4"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.121832 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.179647 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" podStartSLOduration=3.858430674 podStartE2EDuration="1m1.179617452s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.177018402 +0000 UTC m=+1219.536786919" lastFinishedPulling="2026-01-22 10:03:13.49820518 +0000 UTC m=+1276.857973697" observedRunningTime="2026-01-22 10:03:15.1654645 +0000 UTC m=+1278.525233027" watchObservedRunningTime="2026-01-22 10:03:15.179617452 +0000 UTC m=+1278.539385999" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.181310 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" event={"ID":"45ead174-1e73-410e-a2af-5db3a287e6ef","Type":"ContainerStarted","Data":"8ddbfb1c9259a7af2ad8f2ad202326fa631cc4ae8c357cf4dcaf9ff58a17b5a9"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.182020 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.221350 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" podStartSLOduration=4.882063523 podStartE2EDuration="1m2.22132735s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.170205807 +0000 UTC m=+1219.529974324" lastFinishedPulling="2026-01-22 10:03:13.509469624 +0000 UTC m=+1276.869238151" observedRunningTime="2026-01-22 10:03:15.202592193 +0000 UTC m=+1278.562360720" watchObservedRunningTime="2026-01-22 10:03:15.22132735 +0000 UTC m=+1278.581095867" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.231354 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" event={"ID":"fddde490-1908-48b2-afbc-d09bb4ec8e4a","Type":"ContainerStarted","Data":"593bcfc50a7849898cc4962f64f8f71ac022776fc9bb2f1010b611ba3b042fc0"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.231400 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" event={"ID":"fddde490-1908-48b2-afbc-d09bb4ec8e4a","Type":"ContainerStarted","Data":"cb9d40e3fc0c37d62786bc01102ba2a1533dd1b79ce56102c26c3365ca29c555"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.232009 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.253312 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" podStartSLOduration=34.371193814 podStartE2EDuration="1m2.253292744s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.129405362 +0000 UTC m=+1219.489173879" lastFinishedPulling="2026-01-22 10:02:44.011504292 +0000 UTC m=+1247.371272809" observedRunningTime="2026-01-22 10:03:15.234242179 +0000 UTC m=+1278.594010696" watchObservedRunningTime="2026-01-22 10:03:15.253292744 +0000 UTC m=+1278.613061261" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.511358 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" event={"ID":"62432f9a-3dc6-4f17-844e-fd5a6e983097","Type":"ContainerStarted","Data":"f7cad1757eb343a44ee10b634c193086aec54d4cc7f22d9038724269cd250707"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.512290 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.514783 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" podStartSLOduration=4.486593244 podStartE2EDuration="1m1.514771772s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.51542493 +0000 UTC m=+1219.875193437" lastFinishedPulling="2026-01-22 10:03:13.543603438 +0000 UTC m=+1276.903371965" observedRunningTime="2026-01-22 10:03:15.512651025 +0000 UTC m=+1278.872419562" watchObservedRunningTime="2026-01-22 10:03:15.514771772 +0000 UTC m=+1278.874540289" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.539017 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" event={"ID":"bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b","Type":"ContainerStarted","Data":"865a441b525fd833c032e690533aa650bd5e5a68f5a4e0d4435ebbb970db04f9"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.539667 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.573383 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" event={"ID":"a89f20bd-9744-47b0-abcb-07b75b44d255","Type":"ContainerStarted","Data":"e305e1f6d59c18af6531aa83501dfda001fbbfb6440efadb95fca34ac8557311"} Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.573452 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.585474 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" podStartSLOduration=4.654565664 podStartE2EDuration="1m1.585451123s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.509452577 +0000 UTC m=+1219.869221084" lastFinishedPulling="2026-01-22 10:03:13.440338016 +0000 UTC m=+1276.800106543" observedRunningTime="2026-01-22 10:03:15.578404132 +0000 UTC m=+1278.938172649" watchObservedRunningTime="2026-01-22 10:03:15.585451123 +0000 UTC m=+1278.945219640" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.631719 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" podStartSLOduration=4.001564259 podStartE2EDuration="1m2.631697373s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:15.641820003 +0000 UTC m=+1219.001588520" lastFinishedPulling="2026-01-22 10:03:14.271953117 +0000 UTC m=+1277.631721634" observedRunningTime="2026-01-22 10:03:15.615476434 +0000 UTC m=+1278.975244971" watchObservedRunningTime="2026-01-22 10:03:15.631697373 +0000 UTC m=+1278.991465890" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.736934 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" podStartSLOduration=33.974937711 podStartE2EDuration="1m1.736915347s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.249826474 +0000 UTC m=+1219.609594991" lastFinishedPulling="2026-01-22 10:02:44.01180411 +0000 UTC m=+1247.371572627" observedRunningTime="2026-01-22 10:03:15.730447222 +0000 UTC m=+1279.090215739" watchObservedRunningTime="2026-01-22 10:03:15.736915347 +0000 UTC m=+1279.096683864" Jan 22 10:03:15 crc kubenswrapper[4836]: I0122 10:03:15.737167 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" podStartSLOduration=5.814355624 podStartE2EDuration="1m2.737163014s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.517478395 +0000 UTC m=+1219.877246912" lastFinishedPulling="2026-01-22 10:03:13.440285785 +0000 UTC m=+1276.800054302" observedRunningTime="2026-01-22 10:03:15.699542096 +0000 UTC m=+1279.059310613" watchObservedRunningTime="2026-01-22 10:03:15.737163014 +0000 UTC m=+1279.096931531" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.636634 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" event={"ID":"4f4c9bb1-0e84-4d29-9da3-c068c13723db","Type":"ContainerStarted","Data":"9882eb78262743c56579271b31ece3c849c2dfd93825145fcc07f720d901550a"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.637898 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.640642 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" event={"ID":"6ede4b10-e9d1-44fe-a400-4bb900e2dc59","Type":"ContainerStarted","Data":"bcaa54256f6410a2a18b77aa419331e1253ea293d85031dd4a2b51fbf7d1e1bf"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.641139 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.655906 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" event={"ID":"f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7","Type":"ContainerStarted","Data":"fc4c6fdb198c51acaef337d363e70a1541a22d320a48b382f1d05d74cda39f80"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.656550 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.659154 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" event={"ID":"d9562ef6-7f4b-49b3-bd82-eba3360ab2b3","Type":"ContainerStarted","Data":"df5b28a02ebc9041ed6cea0b4b68146f76b25711517932836018df62cbea8ee9"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.659519 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.660399 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" event={"ID":"72203b24-67d7-4dac-8430-51c940eaa7e8","Type":"ContainerStarted","Data":"c2bf5567d508341707fc5cefc8ff561fe3428f29bf107bc2b227ccb2eeecd87f"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.660795 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.668216 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" podStartSLOduration=5.543959067 podStartE2EDuration="1m3.668196501s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:15.372461246 +0000 UTC m=+1218.732229763" lastFinishedPulling="2026-01-22 10:03:13.49669868 +0000 UTC m=+1276.856467197" observedRunningTime="2026-01-22 10:03:16.658852679 +0000 UTC m=+1280.018621196" watchObservedRunningTime="2026-01-22 10:03:16.668196501 +0000 UTC m=+1280.027965018" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.668425 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" podStartSLOduration=62.668419037 podStartE2EDuration="1m2.668419037s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:03:15.817451984 +0000 UTC m=+1279.177220501" watchObservedRunningTime="2026-01-22 10:03:16.668419037 +0000 UTC m=+1280.028187554" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.670733 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" event={"ID":"84ad66e0-8ccb-4bd1-a5c6-ec82697798ed","Type":"ContainerStarted","Data":"12dcb5245e7a7b061cda3348cf3d91b7d3b10f65c838dfa81fa039b01efc532c"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.671475 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.676441 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" event={"ID":"8a4951f8-3310-4e38-b989-fa900a0e68bd","Type":"ContainerStarted","Data":"8b8cea555782a42c73f800f196eff9c58cd0d73b574dc4d7ebb939f11928bc73"} Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.682324 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" podStartSLOduration=4.683667213 podStartE2EDuration="1m2.682312213s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.191581336 +0000 UTC m=+1219.551349863" lastFinishedPulling="2026-01-22 10:03:14.190226346 +0000 UTC m=+1277.549994863" observedRunningTime="2026-01-22 10:03:16.68107375 +0000 UTC m=+1280.040842277" watchObservedRunningTime="2026-01-22 10:03:16.682312213 +0000 UTC m=+1280.042080730" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.736863 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" podStartSLOduration=5.200593595 podStartE2EDuration="1m2.736846337s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.174567085 +0000 UTC m=+1219.534335602" lastFinishedPulling="2026-01-22 10:03:13.710819827 +0000 UTC m=+1277.070588344" observedRunningTime="2026-01-22 10:03:16.736322702 +0000 UTC m=+1280.096091209" watchObservedRunningTime="2026-01-22 10:03:16.736846337 +0000 UTC m=+1280.096614844" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.870016 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" podStartSLOduration=5.850449698 podStartE2EDuration="1m2.869997766s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.475543858 +0000 UTC m=+1219.835312375" lastFinishedPulling="2026-01-22 10:03:13.495091926 +0000 UTC m=+1276.854860443" observedRunningTime="2026-01-22 10:03:16.869073821 +0000 UTC m=+1280.228842328" watchObservedRunningTime="2026-01-22 10:03:16.869997766 +0000 UTC m=+1280.229766283" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.875842 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-gkb6r" podStartSLOduration=5.853281987 podStartE2EDuration="1m2.875824683s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.521097183 +0000 UTC m=+1219.880865700" lastFinishedPulling="2026-01-22 10:03:13.543639879 +0000 UTC m=+1276.903408396" observedRunningTime="2026-01-22 10:03:16.83313175 +0000 UTC m=+1280.192900267" watchObservedRunningTime="2026-01-22 10:03:16.875824683 +0000 UTC m=+1280.235593200" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.956326 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" podStartSLOduration=5.9980957230000005 podStartE2EDuration="1m3.956309149s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.232734751 +0000 UTC m=+1219.592503268" lastFinishedPulling="2026-01-22 10:03:14.190948177 +0000 UTC m=+1277.550716694" observedRunningTime="2026-01-22 10:03:16.915488235 +0000 UTC m=+1280.275256752" watchObservedRunningTime="2026-01-22 10:03:16.956309149 +0000 UTC m=+1280.316077656" Jan 22 10:03:16 crc kubenswrapper[4836]: I0122 10:03:16.957082 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" podStartSLOduration=6.362702118 podStartE2EDuration="1m3.95706635s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:02:16.177378142 +0000 UTC m=+1219.537146659" lastFinishedPulling="2026-01-22 10:03:13.771742374 +0000 UTC m=+1277.131510891" observedRunningTime="2026-01-22 10:03:16.950579795 +0000 UTC m=+1280.310348312" watchObservedRunningTime="2026-01-22 10:03:16.95706635 +0000 UTC m=+1280.316834867" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.282119 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-7dnbq" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.323663 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-dlkpl" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.337181 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-hpfrj" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.360879 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-zxzq5" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.404650 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-glr2h" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.421120 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-9v24b" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.421423 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-hfd9z" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.486495 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-gh4kf" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.497655 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-vf8kq" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.524720 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-h5xn2" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.553437 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-4547l" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.629777 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-s4pbs" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.743441 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-pgqb5" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.770586 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-vxfhs" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.817722 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-2dctq" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.857367 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-vq98j" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.888097 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-rtzb5" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.902034 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-628hc" Jan 22 10:03:24 crc kubenswrapper[4836]: I0122 10:03:24.984497 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-8h7wm" Jan 22 10:03:26 crc kubenswrapper[4836]: I0122 10:03:26.745355 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-jtvkz" Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.822082 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" event={"ID":"43c5f5ad-e7e7-460d-9ed2-555cc840253f","Type":"ContainerStarted","Data":"158a40ae26762f15b34d0f6e3297d18d6e8677ca010e312618969219a8dcf25e"} Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.822526 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.824355 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" event={"ID":"19f29c0b-c9b6-4137-9558-6af7128aa0b9","Type":"ContainerStarted","Data":"24947517ff403befdd957a68f067ed1e81ab8b97119fc06907f970b93b766e6f"} Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.824529 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.869671 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" podStartSLOduration=61.954193179 podStartE2EDuration="1m15.869652227s" podCreationTimestamp="2026-01-22 10:02:13 +0000 UTC" firstStartedPulling="2026-01-22 10:03:14.071887458 +0000 UTC m=+1277.431655975" lastFinishedPulling="2026-01-22 10:03:27.987346506 +0000 UTC m=+1291.347115023" observedRunningTime="2026-01-22 10:03:28.867099788 +0000 UTC m=+1292.226868305" watchObservedRunningTime="2026-01-22 10:03:28.869652227 +0000 UTC m=+1292.229420744" Jan 22 10:03:28 crc kubenswrapper[4836]: I0122 10:03:28.908174 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" podStartSLOduration=61.016739088 podStartE2EDuration="1m14.908158187s" podCreationTimestamp="2026-01-22 10:02:14 +0000 UTC" firstStartedPulling="2026-01-22 10:03:14.077607512 +0000 UTC m=+1277.437376029" lastFinishedPulling="2026-01-22 10:03:27.969026611 +0000 UTC m=+1291.328795128" observedRunningTime="2026-01-22 10:03:28.905662619 +0000 UTC m=+1292.265431136" watchObservedRunningTime="2026-01-22 10:03:28.908158187 +0000 UTC m=+1292.267926714" Jan 22 10:03:36 crc kubenswrapper[4836]: I0122 10:03:36.257571 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-4p2bd" Jan 22 10:03:36 crc kubenswrapper[4836]: I0122 10:03:36.641593 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs" Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.859993 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-j5j4k/must-gather-6dcfr"] Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.861915 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.865935 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j5j4k"/"kube-root-ca.crt" Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.868341 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-j5j4k"/"openshift-service-ca.crt" Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.881799 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j5j4k/must-gather-6dcfr"] Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.959791 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n675v\" (UniqueName: \"kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:35 crc kubenswrapper[4836]: I0122 10:04:35.959865 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.061354 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n675v\" (UniqueName: \"kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.061423 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.061903 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.081926 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n675v\" (UniqueName: \"kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v\") pod \"must-gather-6dcfr\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.181734 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:04:36 crc kubenswrapper[4836]: I0122 10:04:36.417465 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-j5j4k/must-gather-6dcfr"] Jan 22 10:04:37 crc kubenswrapper[4836]: I0122 10:04:37.371230 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" event={"ID":"f0f02dd2-e3a1-4c0e-9b0c-22385660521f","Type":"ContainerStarted","Data":"dcd2ea441644f8d0ae4da37b3f4d59c75b3dfbc6f41f7cd8f14b617664b69dcb"} Jan 22 10:04:48 crc kubenswrapper[4836]: I0122 10:04:48.460599 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" event={"ID":"f0f02dd2-e3a1-4c0e-9b0c-22385660521f","Type":"ContainerStarted","Data":"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623"} Jan 22 10:04:49 crc kubenswrapper[4836]: I0122 10:04:49.468526 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" event={"ID":"f0f02dd2-e3a1-4c0e-9b0c-22385660521f","Type":"ContainerStarted","Data":"de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54"} Jan 22 10:04:49 crc kubenswrapper[4836]: I0122 10:04:49.485855 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" podStartSLOduration=2.81369272 podStartE2EDuration="14.485834699s" podCreationTimestamp="2026-01-22 10:04:35 +0000 UTC" firstStartedPulling="2026-01-22 10:04:36.425144823 +0000 UTC m=+1359.784913340" lastFinishedPulling="2026-01-22 10:04:48.097286792 +0000 UTC m=+1371.457055319" observedRunningTime="2026-01-22 10:04:49.481848241 +0000 UTC m=+1372.841616768" watchObservedRunningTime="2026-01-22 10:04:49.485834699 +0000 UTC m=+1372.845603226" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.542892 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-7dnbq_bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.557265 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-dlkpl_e5ac1213-b336-49f2-a1c3-93fbbbdc724c/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.571770 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-hpfrj_4f4c9bb1-0e84-4d29-9da3-c068c13723db/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.584845 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/extract/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.593297 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/util/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.604802 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/pull/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.619654 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-zxzq5_45ead174-1e73-410e-a2af-5db3a287e6ef/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.630528 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-glr2h_2d3d18eb-0f1a-4552-91d3-6c6bec689f12/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.642879 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-hfd9z_62432f9a-3dc6-4f17-844e-fd5a6e983097/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.658167 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-4p2bd_43c5f5ad-e7e7-460d-9ed2-555cc840253f/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.671237 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-gh4kf_72203b24-67d7-4dac-8430-51c940eaa7e8/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.680525 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-vf8kq_84ad66e0-8ccb-4bd1-a5c6-ec82697798ed/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.693961 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-h5xn2_4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.708835 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-4547l_ab605e98-a012-403e-9369-aa87d02895ba/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.718483 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-9v24b_f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.737281 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-pgqb5_a89f20bd-9744-47b0-abcb-07b75b44d255/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.749207 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-vxfhs_2e57fccc-cded-48d6-bd9a-c797d2013d6d/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.765483 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs_19f29c0b-c9b6-4137-9558-6af7128aa0b9/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.917487 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-69wfj_1782e8eb-b211-494d-8091-45ef1630b6c6/operator/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.933704 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-jtvkz_fddde490-1908-48b2-afbc-d09bb4ec8e4a/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.949651 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vsgjh_46e9eb65-0aaa-45dc-810a-7c0e8cc089a5/registry-server/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.958263 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-2dctq_0fbc160a-787a-4a04-ab60-ddfa07e6d6ca/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.968217 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-rtzb5_d9562ef6-7f4b-49b3-bd82-eba3360ab2b3/manager/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.980779 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-gkb6r_8a4951f8-3310-4e38-b989-fa900a0e68bd/operator/0.log" Jan 22 10:04:55 crc kubenswrapper[4836]: I0122 10:04:55.993622 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-s4pbs_6ede4b10-e9d1-44fe-a400-4bb900e2dc59/manager/0.log" Jan 22 10:04:56 crc kubenswrapper[4836]: I0122 10:04:56.008300 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-8h7wm_34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed/manager/0.log" Jan 22 10:04:56 crc kubenswrapper[4836]: I0122 10:04:56.018019 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-vq98j_ada6a395-b944-4798-bb31-347fbc476be4/manager/0.log" Jan 22 10:04:56 crc kubenswrapper[4836]: I0122 10:04:56.031635 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-628hc_155c2cd6-e900-4e96-bc0b-2b3227732430/manager/0.log" Jan 22 10:05:03 crc kubenswrapper[4836]: I0122 10:05:03.686778 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-stt4s_92cea25b-5193-412c-bd80-eb4932e7a7a5/control-plane-machine-set-operator/0.log" Jan 22 10:05:03 crc kubenswrapper[4836]: I0122 10:05:03.705009 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btclw_1121b7db-f2c3-4900-81c7-e765a19a091c/kube-rbac-proxy/0.log" Jan 22 10:05:03 crc kubenswrapper[4836]: I0122 10:05:03.724348 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btclw_1121b7db-f2c3-4900-81c7-e765a19a091c/machine-api-operator/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.660883 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/controller/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.666403 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/kube-rbac-proxy/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.683258 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/controller/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.879808 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.888158 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/reloader/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.892370 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr-metrics/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.898138 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.903181 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy-frr/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.909344 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-frr-files/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.915617 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-reloader/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.921392 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-metrics/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.932581 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-4lq59_c11f6970-4b6e-4574-b9a6-0d50617198dc/frr-k8s-webhook-server/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.957080 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-799d7d6dcc-bvjvp_acf969fa-26d3-4c92-9b0e-553cedfd3a04/manager/0.log" Jan 22 10:05:12 crc kubenswrapper[4836]: I0122 10:05:12.965590 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7fd664ffb8-qplwn_e785c5eb-f7d6-4c65-8632-16a4d7e4f161/webhook-server/0.log" Jan 22 10:05:13 crc kubenswrapper[4836]: I0122 10:05:13.129793 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/speaker/0.log" Jan 22 10:05:13 crc kubenswrapper[4836]: I0122 10:05:13.134729 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/kube-rbac-proxy/0.log" Jan 22 10:05:16 crc kubenswrapper[4836]: I0122 10:05:16.110622 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-4qvrz_dde5f84b-448c-4059-9adb-e5858090d8c4/cert-manager-controller/0.log" Jan 22 10:05:16 crc kubenswrapper[4836]: I0122 10:05:16.127701 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-jhm6c_9994d233-12a4-49af-8e54-b76382ca5efb/cert-manager-cainjector/0.log" Jan 22 10:05:16 crc kubenswrapper[4836]: I0122 10:05:16.143090 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-b8jpq_b47c9aca-4964-4a7f-8588-622b55539ec8/cert-manager-webhook/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.865035 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-l29ws_e97bdfd0-897c-45ac-9597-4d8c24e9d279/nmstate-console-plugin/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.880159 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-59bdc_63fcf02d-2d6a-4e47-8725-a4dde8fac447/nmstate-handler/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.894142 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-c8v9m_71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9/nmstate-metrics/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.905413 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-c8v9m_71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9/kube-rbac-proxy/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.919694 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-nwcbn_e6db6563-9866-4671-9f24-802e8717b5d7/nmstate-operator/0.log" Jan 22 10:05:20 crc kubenswrapper[4836]: I0122 10:05:20.931356 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-xw45j_b89c6d12-f385-4e08-9a29-43befa6f48cc/nmstate-webhook/0.log" Jan 22 10:05:23 crc kubenswrapper[4836]: I0122 10:05:23.897859 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:05:23 crc kubenswrapper[4836]: I0122 10:05:23.898217 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.012799 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/controller/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.020605 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/kube-rbac-proxy/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.038327 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/controller/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.217719 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.226591 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/reloader/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.232926 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr-metrics/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.239229 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.245717 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy-frr/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.253357 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-frr-files/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.259311 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-reloader/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.266518 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-metrics/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.275471 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-4lq59_c11f6970-4b6e-4574-b9a6-0d50617198dc/frr-k8s-webhook-server/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.301368 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-799d7d6dcc-bvjvp_acf969fa-26d3-4c92-9b0e-553cedfd3a04/manager/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.317538 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7fd664ffb8-qplwn_e785c5eb-f7d6-4c65-8632-16a4d7e4f161/webhook-server/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.464341 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/speaker/0.log" Jan 22 10:05:31 crc kubenswrapper[4836]: I0122 10:05:31.471344 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/kube-rbac-proxy/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.032283 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm_3b5f2dd9-4392-4345-9241-b0b3a3cc2936/extract/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.041010 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm_3b5f2dd9-4392-4345-9241-b0b3a3cc2936/util/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.078859 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ann5mm_3b5f2dd9-4392-4345-9241-b0b3a3cc2936/pull/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.088843 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk_3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd/extract/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.094104 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk_3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd/util/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.103776 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcsvwbk_3d00ac00-ae92-4924-b5cd-bcd7ee7b2acd/pull/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.113518 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2_37685439-5383-47b3-8cd0-ad982b0a7b5e/extract/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.121392 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2_37685439-5383-47b3-8cd0-ad982b0a7b5e/util/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.130802 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713kx6h2_37685439-5383-47b3-8cd0-ad982b0a7b5e/pull/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.345711 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bntn_dfab7eb6-8f9f-44cd-8cc4-2e364698c767/registry-server/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.352887 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bntn_dfab7eb6-8f9f-44cd-8cc4-2e364698c767/extract-utilities/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.363706 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9bntn_dfab7eb6-8f9f-44cd-8cc4-2e364698c767/extract-content/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.576522 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z48bp_8d7f3c8f-bc53-4fb9-948a-408169582191/registry-server/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.581151 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z48bp_8d7f3c8f-bc53-4fb9-948a-408169582191/extract-utilities/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.592629 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z48bp_8d7f3c8f-bc53-4fb9-948a-408169582191/extract-content/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.610498 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-crc7k_66077fb8-1811-4d25-8b81-9ca672521135/marketplace-operator/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.681941 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m67hn_d131c247-f421-4d22-8e20-4605e2e6cd18/registry-server/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.688111 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m67hn_d131c247-f421-4d22-8e20-4605e2e6cd18/extract-utilities/0.log" Jan 22 10:05:36 crc kubenswrapper[4836]: I0122 10:05:36.694903 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m67hn_d131c247-f421-4d22-8e20-4605e2e6cd18/extract-content/0.log" Jan 22 10:05:37 crc kubenswrapper[4836]: I0122 10:05:37.033742 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrlz2_a93e2393-3439-4b33-ac03-142faa9de7de/registry-server/0.log" Jan 22 10:05:37 crc kubenswrapper[4836]: I0122 10:05:37.039415 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrlz2_a93e2393-3439-4b33-ac03-142faa9de7de/extract-utilities/0.log" Jan 22 10:05:37 crc kubenswrapper[4836]: I0122 10:05:37.047723 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrlz2_a93e2393-3439-4b33-ac03-142faa9de7de/extract-content/0.log" Jan 22 10:05:53 crc kubenswrapper[4836]: I0122 10:05:53.897968 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:05:53 crc kubenswrapper[4836]: I0122 10:05:53.898430 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.191906 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/controller/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.197948 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-phjc8_9437c57d-62ec-44b0-bf05-2e42c76e6f49/kube-rbac-proxy/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.217486 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/controller/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.291986 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-4qvrz_dde5f84b-448c-4059-9adb-e5858090d8c4/cert-manager-controller/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.308654 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-jhm6c_9994d233-12a4-49af-8e54-b76382ca5efb/cert-manager-cainjector/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.322583 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-b8jpq_b47c9aca-4964-4a7f-8588-622b55539ec8/cert-manager-webhook/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.392773 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.399770 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/reloader/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.406365 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/frr-metrics/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.418814 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.424886 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/kube-rbac-proxy-frr/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.435502 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-frr-files/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.442887 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-reloader/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.452600 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pvr6q_f3969312-f604-49ff-9560-a6078e253ee9/cp-metrics/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.471124 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-4lq59_c11f6970-4b6e-4574-b9a6-0d50617198dc/frr-k8s-webhook-server/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.490325 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-799d7d6dcc-bvjvp_acf969fa-26d3-4c92-9b0e-553cedfd3a04/manager/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.505529 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7fd664ffb8-qplwn_e785c5eb-f7d6-4c65-8632-16a4d7e4f161/webhook-server/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.640789 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/speaker/0.log" Jan 22 10:06:16 crc kubenswrapper[4836]: I0122 10:06:16.648172 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-v5n2k_e1d933f7-c1d1-49d0-8b0d-d0df7ec9167c/kube-rbac-proxy/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.203662 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-7dnbq_bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.217608 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-dlkpl_e5ac1213-b336-49f2-a1c3-93fbbbdc724c/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.229563 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-hpfrj_4f4c9bb1-0e84-4d29-9da3-c068c13723db/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.238623 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/extract/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.247546 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/util/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.255692 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/pull/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.272631 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-zxzq5_45ead174-1e73-410e-a2af-5db3a287e6ef/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.284571 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-glr2h_2d3d18eb-0f1a-4552-91d3-6c6bec689f12/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.298546 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-hfd9z_62432f9a-3dc6-4f17-844e-fd5a6e983097/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.316552 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-4p2bd_43c5f5ad-e7e7-460d-9ed2-555cc840253f/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.329221 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-gh4kf_72203b24-67d7-4dac-8430-51c940eaa7e8/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.341528 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-vf8kq_84ad66e0-8ccb-4bd1-a5c6-ec82697798ed/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.352936 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-h5xn2_4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.364497 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-4547l_ab605e98-a012-403e-9369-aa87d02895ba/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.376723 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-9v24b_f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.391507 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-pgqb5_a89f20bd-9744-47b0-abcb-07b75b44d255/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.403229 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-vxfhs_2e57fccc-cded-48d6-bd9a-c797d2013d6d/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.419407 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs_19f29c0b-c9b6-4137-9558-6af7128aa0b9/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.533836 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-69wfj_1782e8eb-b211-494d-8091-45ef1630b6c6/operator/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.555382 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-jtvkz_fddde490-1908-48b2-afbc-d09bb4ec8e4a/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.562093 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-4qvrz_dde5f84b-448c-4059-9adb-e5858090d8c4/cert-manager-controller/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.576479 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-jhm6c_9994d233-12a4-49af-8e54-b76382ca5efb/cert-manager-cainjector/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.577147 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vsgjh_46e9eb65-0aaa-45dc-810a-7c0e8cc089a5/registry-server/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.586591 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-b8jpq_b47c9aca-4964-4a7f-8588-622b55539ec8/cert-manager-webhook/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.587545 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-2dctq_0fbc160a-787a-4a04-ab60-ddfa07e6d6ca/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.600478 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-rtzb5_d9562ef6-7f4b-49b3-bd82-eba3360ab2b3/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.613834 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-gkb6r_8a4951f8-3310-4e38-b989-fa900a0e68bd/operator/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.629240 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-s4pbs_6ede4b10-e9d1-44fe-a400-4bb900e2dc59/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.673088 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-8h7wm_34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.683587 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-vq98j_ada6a395-b944-4798-bb31-347fbc476be4/manager/0.log" Jan 22 10:06:17 crc kubenswrapper[4836]: I0122 10:06:17.694055 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-628hc_155c2cd6-e900-4e96-bc0b-2b3227732430/manager/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.261050 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-stt4s_92cea25b-5193-412c-bd80-eb4932e7a7a5/control-plane-machine-set-operator/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.279339 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btclw_1121b7db-f2c3-4900-81c7-e765a19a091c/kube-rbac-proxy/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.288294 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btclw_1121b7db-f2c3-4900-81c7-e765a19a091c/machine-api-operator/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.918100 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-7dnbq_bf2140ed-ce8f-4ed6-b8ea-8251aa1d317b/manager/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.933030 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-dlkpl_e5ac1213-b336-49f2-a1c3-93fbbbdc724c/manager/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.952598 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-hpfrj_4f4c9bb1-0e84-4d29-9da3-c068c13723db/manager/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.965914 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/extract/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.976759 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/util/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.983956 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f9fc79m_76f6be8d-536e-4a8e-ba41-eb2dd7cf75b8/pull/0.log" Jan 22 10:06:18 crc kubenswrapper[4836]: I0122 10:06:18.994623 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-zxzq5_45ead174-1e73-410e-a2af-5db3a287e6ef/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.007956 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-glr2h_2d3d18eb-0f1a-4552-91d3-6c6bec689f12/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.017916 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-hfd9z_62432f9a-3dc6-4f17-844e-fd5a6e983097/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.037642 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-4p2bd_43c5f5ad-e7e7-460d-9ed2-555cc840253f/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.049286 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-gh4kf_72203b24-67d7-4dac-8430-51c940eaa7e8/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.060954 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-vf8kq_84ad66e0-8ccb-4bd1-a5c6-ec82697798ed/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.075597 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-h5xn2_4cdbaa6a-2b38-4350-87c9-e45bf3cc89cf/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.083087 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-4547l_ab605e98-a012-403e-9369-aa87d02895ba/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.091432 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-l29ws_e97bdfd0-897c-45ac-9597-4d8c24e9d279/nmstate-console-plugin/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.093209 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-9v24b_f8cc0fdf-fd8b-41b2-a2af-ef0a958a11b7/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.103427 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-pgqb5_a89f20bd-9744-47b0-abcb-07b75b44d255/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.108867 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-59bdc_63fcf02d-2d6a-4e47-8725-a4dde8fac447/nmstate-handler/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.115871 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-vxfhs_2e57fccc-cded-48d6-bd9a-c797d2013d6d/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.123016 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-c8v9m_71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9/nmstate-metrics/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.130897 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-c8v9m_71b03f9f-75b7-45e3-ac01-fdcc7c5a3fc9/kube-rbac-proxy/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.135614 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b557pbmhs_19f29c0b-c9b6-4137-9558-6af7128aa0b9/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.142840 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-nwcbn_e6db6563-9866-4671-9f24-802e8717b5d7/nmstate-operator/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.154659 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-xw45j_b89c6d12-f385-4e08-9a29-43befa6f48cc/nmstate-webhook/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.253286 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-69wfj_1782e8eb-b211-494d-8091-45ef1630b6c6/operator/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.270602 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-jtvkz_fddde490-1908-48b2-afbc-d09bb4ec8e4a/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.285231 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vsgjh_46e9eb65-0aaa-45dc-810a-7c0e8cc089a5/registry-server/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.295397 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-2dctq_0fbc160a-787a-4a04-ab60-ddfa07e6d6ca/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.305750 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-rtzb5_d9562ef6-7f4b-49b3-bd82-eba3360ab2b3/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.315477 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-gkb6r_8a4951f8-3310-4e38-b989-fa900a0e68bd/operator/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.326166 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-s4pbs_6ede4b10-e9d1-44fe-a400-4bb900e2dc59/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.338063 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-8h7wm_34b2d45c-eab1-4f20-9ddb-fb6769fdd4ed/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.347387 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-vq98j_ada6a395-b944-4798-bb31-347fbc476be4/manager/0.log" Jan 22 10:06:19 crc kubenswrapper[4836]: I0122 10:06:19.360812 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-628hc_155c2cd6-e900-4e96-bc0b-2b3227732430/manager/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.874145 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/3.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.890572 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-89st4_7fcb3e24-f305-4f39-af85-2dc05a0af79f/kube-multus/2.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.900559 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/kube-multus-additional-cni-plugins/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.911463 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/egress-router-binary-copy/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.916393 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/cni-plugins/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.922471 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/bond-cni-plugin/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.928255 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/routeoverride-cni/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.937146 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/whereabouts-cni-bincopy/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.944127 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gfrrz_2aa3f461-19ed-4700-94b6-8c8dd71c6a21/whereabouts-cni/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.958633 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-g5pts_4fe19b48-498b-420a-9a84-9b63a0a1ae9e/multus-admission-controller/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.963158 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-g5pts_4fe19b48-498b-420a-9a84-9b63a0a1ae9e/kube-rbac-proxy/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.981085 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-6kkmw_9ab725c3-0744-400a-b8d7-66e1165476e5/network-metrics-daemon/0.log" Jan 22 10:06:20 crc kubenswrapper[4836]: I0122 10:06:20.985326 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-6kkmw_9ab725c3-0744-400a-b8d7-66e1165476e5/kube-rbac-proxy/0.log" Jan 22 10:06:23 crc kubenswrapper[4836]: I0122 10:06:23.897293 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:06:23 crc kubenswrapper[4836]: I0122 10:06:23.897545 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:06:23 crc kubenswrapper[4836]: I0122 10:06:23.897596 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 10:06:23 crc kubenswrapper[4836]: I0122 10:06:23.898218 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 10:06:23 crc kubenswrapper[4836]: I0122 10:06:23.898266 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62" gracePeriod=600 Jan 22 10:06:24 crc kubenswrapper[4836]: I0122 10:06:24.699031 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62" exitCode=0 Jan 22 10:06:24 crc kubenswrapper[4836]: I0122 10:06:24.699127 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62"} Jan 22 10:06:24 crc kubenswrapper[4836]: I0122 10:06:24.699737 4836 scope.go:117] "RemoveContainer" containerID="98ebf9024b64453dcede5fa4f11620ad5763726f76217cd1fc3ae177392c2067" Jan 22 10:06:25 crc kubenswrapper[4836]: I0122 10:06:25.718010 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a"} Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.608023 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.611288 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.650029 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.717443 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6bxr\" (UniqueName: \"kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.717500 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.717520 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.818965 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6bxr\" (UniqueName: \"kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.819416 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.819527 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.820031 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.820095 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.852221 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6bxr\" (UniqueName: \"kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr\") pod \"redhat-marketplace-9c5kf\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:37 crc kubenswrapper[4836]: I0122 10:08:37.936308 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:38 crc kubenswrapper[4836]: I0122 10:08:38.470422 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:38 crc kubenswrapper[4836]: I0122 10:08:38.765229 4836 generic.go:334] "Generic (PLEG): container finished" podID="92478d54-a686-430c-b2aa-7214f7365052" containerID="172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df" exitCode=0 Jan 22 10:08:38 crc kubenswrapper[4836]: I0122 10:08:38.765400 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerDied","Data":"172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df"} Jan 22 10:08:38 crc kubenswrapper[4836]: I0122 10:08:38.765430 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerStarted","Data":"9752183680cf75578154f1539efe29033ce53dedf769ac73bc931b48b8ec78c1"} Jan 22 10:08:38 crc kubenswrapper[4836]: I0122 10:08:38.766718 4836 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 10:08:40 crc kubenswrapper[4836]: I0122 10:08:40.785373 4836 generic.go:334] "Generic (PLEG): container finished" podID="92478d54-a686-430c-b2aa-7214f7365052" containerID="316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a" exitCode=0 Jan 22 10:08:40 crc kubenswrapper[4836]: I0122 10:08:40.785492 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerDied","Data":"316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a"} Jan 22 10:08:42 crc kubenswrapper[4836]: I0122 10:08:42.812830 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerStarted","Data":"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3"} Jan 22 10:08:42 crc kubenswrapper[4836]: I0122 10:08:42.836459 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9c5kf" podStartSLOduration=2.984576951 podStartE2EDuration="5.836439313s" podCreationTimestamp="2026-01-22 10:08:37 +0000 UTC" firstStartedPulling="2026-01-22 10:08:38.766491589 +0000 UTC m=+1602.126260116" lastFinishedPulling="2026-01-22 10:08:41.618353941 +0000 UTC m=+1604.978122478" observedRunningTime="2026-01-22 10:08:42.829568239 +0000 UTC m=+1606.189336756" watchObservedRunningTime="2026-01-22 10:08:42.836439313 +0000 UTC m=+1606.196207830" Jan 22 10:08:47 crc kubenswrapper[4836]: I0122 10:08:47.937257 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:47 crc kubenswrapper[4836]: I0122 10:08:47.938174 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:47 crc kubenswrapper[4836]: I0122 10:08:47.996540 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:48 crc kubenswrapper[4836]: I0122 10:08:48.901627 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:48 crc kubenswrapper[4836]: I0122 10:08:48.962494 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:50 crc kubenswrapper[4836]: I0122 10:08:50.872158 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9c5kf" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="registry-server" containerID="cri-o://d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3" gracePeriod=2 Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.845830 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.879958 4836 generic.go:334] "Generic (PLEG): container finished" podID="92478d54-a686-430c-b2aa-7214f7365052" containerID="d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3" exitCode=0 Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.880008 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerDied","Data":"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3"} Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.880054 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c5kf" event={"ID":"92478d54-a686-430c-b2aa-7214f7365052","Type":"ContainerDied","Data":"9752183680cf75578154f1539efe29033ce53dedf769ac73bc931b48b8ec78c1"} Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.880395 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c5kf" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.881234 4836 scope.go:117] "RemoveContainer" containerID="d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.899697 4836 scope.go:117] "RemoveContainer" containerID="316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.919527 4836 scope.go:117] "RemoveContainer" containerID="172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.938011 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6bxr\" (UniqueName: \"kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr\") pod \"92478d54-a686-430c-b2aa-7214f7365052\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.938086 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content\") pod \"92478d54-a686-430c-b2aa-7214f7365052\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.938159 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities\") pod \"92478d54-a686-430c-b2aa-7214f7365052\" (UID: \"92478d54-a686-430c-b2aa-7214f7365052\") " Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.939166 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities" (OuterVolumeSpecName: "utilities") pod "92478d54-a686-430c-b2aa-7214f7365052" (UID: "92478d54-a686-430c-b2aa-7214f7365052"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.951212 4836 scope.go:117] "RemoveContainer" containerID="d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.951361 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr" (OuterVolumeSpecName: "kube-api-access-l6bxr") pod "92478d54-a686-430c-b2aa-7214f7365052" (UID: "92478d54-a686-430c-b2aa-7214f7365052"). InnerVolumeSpecName "kube-api-access-l6bxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:08:51 crc kubenswrapper[4836]: E0122 10:08:51.951886 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3\": container with ID starting with d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3 not found: ID does not exist" containerID="d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.951933 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3"} err="failed to get container status \"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3\": rpc error: code = NotFound desc = could not find container \"d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3\": container with ID starting with d6248337c5b45013b21b192fef573d8803020e4933c84513dd45c147f2a9ccb3 not found: ID does not exist" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.951963 4836 scope.go:117] "RemoveContainer" containerID="316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a" Jan 22 10:08:51 crc kubenswrapper[4836]: E0122 10:08:51.952471 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a\": container with ID starting with 316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a not found: ID does not exist" containerID="316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.952545 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a"} err="failed to get container status \"316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a\": rpc error: code = NotFound desc = could not find container \"316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a\": container with ID starting with 316b32f10b3ee2d381eef68eed47d589bdda190df246c25cb65aa680e0d9ef0a not found: ID does not exist" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.952581 4836 scope.go:117] "RemoveContainer" containerID="172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df" Jan 22 10:08:51 crc kubenswrapper[4836]: E0122 10:08:51.952885 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df\": container with ID starting with 172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df not found: ID does not exist" containerID="172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.952908 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df"} err="failed to get container status \"172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df\": rpc error: code = NotFound desc = could not find container \"172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df\": container with ID starting with 172a81f6d5c018b145154a06f79ca7bbb1e45229c03582f6b41ff3119a4e84df not found: ID does not exist" Jan 22 10:08:51 crc kubenswrapper[4836]: I0122 10:08:51.964458 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92478d54-a686-430c-b2aa-7214f7365052" (UID: "92478d54-a686-430c-b2aa-7214f7365052"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:08:52 crc kubenswrapper[4836]: I0122 10:08:52.039644 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6bxr\" (UniqueName: \"kubernetes.io/projected/92478d54-a686-430c-b2aa-7214f7365052-kube-api-access-l6bxr\") on node \"crc\" DevicePath \"\"" Jan 22 10:08:52 crc kubenswrapper[4836]: I0122 10:08:52.039682 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:08:52 crc kubenswrapper[4836]: I0122 10:08:52.039696 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92478d54-a686-430c-b2aa-7214f7365052-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:08:52 crc kubenswrapper[4836]: I0122 10:08:52.229572 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:52 crc kubenswrapper[4836]: I0122 10:08:52.237817 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c5kf"] Jan 22 10:08:53 crc kubenswrapper[4836]: I0122 10:08:53.007155 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92478d54-a686-430c-b2aa-7214f7365052" path="/var/lib/kubelet/pods/92478d54-a686-430c-b2aa-7214f7365052/volumes" Jan 22 10:08:53 crc kubenswrapper[4836]: I0122 10:08:53.897227 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:08:53 crc kubenswrapper[4836]: I0122 10:08:53.897314 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.144387 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:07 crc kubenswrapper[4836]: E0122 10:09:07.145517 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="extract-utilities" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.145539 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="extract-utilities" Jan 22 10:09:07 crc kubenswrapper[4836]: E0122 10:09:07.145553 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="registry-server" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.145562 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="registry-server" Jan 22 10:09:07 crc kubenswrapper[4836]: E0122 10:09:07.145578 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="extract-content" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.145587 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="extract-content" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.145757 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="92478d54-a686-430c-b2aa-7214f7365052" containerName="registry-server" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.147226 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.159668 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.286815 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-225bc\" (UniqueName: \"kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.286870 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.286960 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.388201 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.388275 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-225bc\" (UniqueName: \"kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.388303 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.388667 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.388871 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.405406 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-225bc\" (UniqueName: \"kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc\") pod \"community-operators-vtr9c\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.471162 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:07 crc kubenswrapper[4836]: I0122 10:09:07.969538 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:07 crc kubenswrapper[4836]: W0122 10:09:07.972814 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f6d0cd8_1f48_455a_b45f_ce41a924c15e.slice/crio-0352a9af45dfdee75fd5cc1a52cd022c8f74ed664d9496e125f410473d5346e2 WatchSource:0}: Error finding container 0352a9af45dfdee75fd5cc1a52cd022c8f74ed664d9496e125f410473d5346e2: Status 404 returned error can't find the container with id 0352a9af45dfdee75fd5cc1a52cd022c8f74ed664d9496e125f410473d5346e2 Jan 22 10:09:08 crc kubenswrapper[4836]: I0122 10:09:08.010986 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerStarted","Data":"0352a9af45dfdee75fd5cc1a52cd022c8f74ed664d9496e125f410473d5346e2"} Jan 22 10:09:09 crc kubenswrapper[4836]: I0122 10:09:09.019905 4836 generic.go:334] "Generic (PLEG): container finished" podID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerID="8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a" exitCode=0 Jan 22 10:09:09 crc kubenswrapper[4836]: I0122 10:09:09.019985 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerDied","Data":"8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a"} Jan 22 10:09:10 crc kubenswrapper[4836]: I0122 10:09:10.029364 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerStarted","Data":"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60"} Jan 22 10:09:11 crc kubenswrapper[4836]: I0122 10:09:11.037815 4836 generic.go:334] "Generic (PLEG): container finished" podID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerID="5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60" exitCode=0 Jan 22 10:09:11 crc kubenswrapper[4836]: I0122 10:09:11.037881 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerDied","Data":"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60"} Jan 22 10:09:12 crc kubenswrapper[4836]: I0122 10:09:12.049126 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerStarted","Data":"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811"} Jan 22 10:09:12 crc kubenswrapper[4836]: I0122 10:09:12.075562 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vtr9c" podStartSLOduration=2.64222278 podStartE2EDuration="5.075527575s" podCreationTimestamp="2026-01-22 10:09:07 +0000 UTC" firstStartedPulling="2026-01-22 10:09:09.02254315 +0000 UTC m=+1632.382311667" lastFinishedPulling="2026-01-22 10:09:11.455847925 +0000 UTC m=+1634.815616462" observedRunningTime="2026-01-22 10:09:12.072472821 +0000 UTC m=+1635.432241378" watchObservedRunningTime="2026-01-22 10:09:12.075527575 +0000 UTC m=+1635.435296142" Jan 22 10:09:17 crc kubenswrapper[4836]: I0122 10:09:17.471980 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:17 crc kubenswrapper[4836]: I0122 10:09:17.472592 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:17 crc kubenswrapper[4836]: I0122 10:09:17.516340 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:18 crc kubenswrapper[4836]: I0122 10:09:18.163816 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:18 crc kubenswrapper[4836]: I0122 10:09:18.229389 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.111271 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vtr9c" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="registry-server" containerID="cri-o://e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811" gracePeriod=2 Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.531453 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.699673 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities\") pod \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.700067 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-225bc\" (UniqueName: \"kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc\") pod \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.700297 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content\") pod \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\" (UID: \"5f6d0cd8-1f48-455a-b45f-ce41a924c15e\") " Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.700980 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities" (OuterVolumeSpecName: "utilities") pod "5f6d0cd8-1f48-455a-b45f-ce41a924c15e" (UID: "5f6d0cd8-1f48-455a-b45f-ce41a924c15e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.707285 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc" (OuterVolumeSpecName: "kube-api-access-225bc") pod "5f6d0cd8-1f48-455a-b45f-ce41a924c15e" (UID: "5f6d0cd8-1f48-455a-b45f-ce41a924c15e"). InnerVolumeSpecName "kube-api-access-225bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.764163 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f6d0cd8-1f48-455a-b45f-ce41a924c15e" (UID: "5f6d0cd8-1f48-455a-b45f-ce41a924c15e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.802204 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.802249 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-225bc\" (UniqueName: \"kubernetes.io/projected/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-kube-api-access-225bc\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:20 crc kubenswrapper[4836]: I0122 10:09:20.802263 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f6d0cd8-1f48-455a-b45f-ce41a924c15e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.122340 4836 generic.go:334] "Generic (PLEG): container finished" podID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerID="e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811" exitCode=0 Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.122393 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerDied","Data":"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811"} Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.122472 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vtr9c" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.122500 4836 scope.go:117] "RemoveContainer" containerID="e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.122483 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vtr9c" event={"ID":"5f6d0cd8-1f48-455a-b45f-ce41a924c15e","Type":"ContainerDied","Data":"0352a9af45dfdee75fd5cc1a52cd022c8f74ed664d9496e125f410473d5346e2"} Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.153957 4836 scope.go:117] "RemoveContainer" containerID="5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.161160 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.169925 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vtr9c"] Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.179863 4836 scope.go:117] "RemoveContainer" containerID="8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.218360 4836 scope.go:117] "RemoveContainer" containerID="e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811" Jan 22 10:09:21 crc kubenswrapper[4836]: E0122 10:09:21.221650 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811\": container with ID starting with e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811 not found: ID does not exist" containerID="e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.221732 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811"} err="failed to get container status \"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811\": rpc error: code = NotFound desc = could not find container \"e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811\": container with ID starting with e0a8d8bedbf43cd5ea96d841234eb5aeea6f34cc09e845bd4700f722a031b811 not found: ID does not exist" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.221791 4836 scope.go:117] "RemoveContainer" containerID="5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60" Jan 22 10:09:21 crc kubenswrapper[4836]: E0122 10:09:21.222328 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60\": container with ID starting with 5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60 not found: ID does not exist" containerID="5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.222458 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60"} err="failed to get container status \"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60\": rpc error: code = NotFound desc = could not find container \"5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60\": container with ID starting with 5c7800e9aeb907d672a32c899bc44b4b7017442bfc2f1e716ac3a7c2b017dd60 not found: ID does not exist" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.222565 4836 scope.go:117] "RemoveContainer" containerID="8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a" Jan 22 10:09:21 crc kubenswrapper[4836]: E0122 10:09:21.222954 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a\": container with ID starting with 8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a not found: ID does not exist" containerID="8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a" Jan 22 10:09:21 crc kubenswrapper[4836]: I0122 10:09:21.223068 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a"} err="failed to get container status \"8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a\": rpc error: code = NotFound desc = could not find container \"8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a\": container with ID starting with 8da62ae118bbb5360ab5c80d081d20836e59ca8187686e3cc0291c433504ba8a not found: ID does not exist" Jan 22 10:09:23 crc kubenswrapper[4836]: I0122 10:09:23.018011 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" path="/var/lib/kubelet/pods/5f6d0cd8-1f48-455a-b45f-ce41a924c15e/volumes" Jan 22 10:09:23 crc kubenswrapper[4836]: I0122 10:09:23.897901 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:09:23 crc kubenswrapper[4836]: I0122 10:09:23.898316 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.327550 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:09:44 crc kubenswrapper[4836]: E0122 10:09:44.336761 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="extract-content" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.337071 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="extract-content" Jan 22 10:09:44 crc kubenswrapper[4836]: E0122 10:09:44.337120 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="extract-utilities" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.337130 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="extract-utilities" Jan 22 10:09:44 crc kubenswrapper[4836]: E0122 10:09:44.337149 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="registry-server" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.337156 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="registry-server" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.337384 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6d0cd8-1f48-455a-b45f-ce41a924c15e" containerName="registry-server" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.338635 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.358812 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.449885 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.449981 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.450011 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgh2w\" (UniqueName: \"kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.551583 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.551643 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgh2w\" (UniqueName: \"kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.551740 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.552211 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.552241 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.570532 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgh2w\" (UniqueName: \"kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w\") pod \"certified-operators-qj4w9\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:44 crc kubenswrapper[4836]: I0122 10:09:44.682931 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:45 crc kubenswrapper[4836]: I0122 10:09:45.179288 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:09:45 crc kubenswrapper[4836]: I0122 10:09:45.358632 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerStarted","Data":"918ac1f98be0c741969453031eb4ab9d5a3d38e048e91bd0361a209d7866ea09"} Jan 22 10:09:46 crc kubenswrapper[4836]: I0122 10:09:46.365747 4836 generic.go:334] "Generic (PLEG): container finished" podID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerID="97bbd6add7cb26a7c2d53a7536546803dff5f26f19b9df6a6d394d97b905c4cb" exitCode=0 Jan 22 10:09:46 crc kubenswrapper[4836]: I0122 10:09:46.365929 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerDied","Data":"97bbd6add7cb26a7c2d53a7536546803dff5f26f19b9df6a6d394d97b905c4cb"} Jan 22 10:09:48 crc kubenswrapper[4836]: I0122 10:09:48.383002 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerStarted","Data":"f71deaf14fa8797f61d62afa9407b5df72f021238fad41e8dede8d2a5656230e"} Jan 22 10:09:49 crc kubenswrapper[4836]: I0122 10:09:49.395642 4836 generic.go:334] "Generic (PLEG): container finished" podID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerID="f71deaf14fa8797f61d62afa9407b5df72f021238fad41e8dede8d2a5656230e" exitCode=0 Jan 22 10:09:49 crc kubenswrapper[4836]: I0122 10:09:49.395933 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerDied","Data":"f71deaf14fa8797f61d62afa9407b5df72f021238fad41e8dede8d2a5656230e"} Jan 22 10:09:50 crc kubenswrapper[4836]: I0122 10:09:50.404538 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerStarted","Data":"1a15a86442c5722afd98083a83ec9bd60dad67fdc7499f77b2cf5dbe09f06ed2"} Jan 22 10:09:50 crc kubenswrapper[4836]: I0122 10:09:50.431261 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qj4w9" podStartSLOduration=2.647116443 podStartE2EDuration="6.431237722s" podCreationTimestamp="2026-01-22 10:09:44 +0000 UTC" firstStartedPulling="2026-01-22 10:09:46.367743073 +0000 UTC m=+1669.727511590" lastFinishedPulling="2026-01-22 10:09:50.151864352 +0000 UTC m=+1673.511632869" observedRunningTime="2026-01-22 10:09:50.423955796 +0000 UTC m=+1673.783724323" watchObservedRunningTime="2026-01-22 10:09:50.431237722 +0000 UTC m=+1673.791006249" Jan 22 10:09:53 crc kubenswrapper[4836]: I0122 10:09:53.897590 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:09:53 crc kubenswrapper[4836]: I0122 10:09:53.898859 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:09:53 crc kubenswrapper[4836]: I0122 10:09:53.899031 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 10:09:53 crc kubenswrapper[4836]: I0122 10:09:53.899810 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 10:09:53 crc kubenswrapper[4836]: I0122 10:09:53.899983 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" gracePeriod=600 Jan 22 10:09:54 crc kubenswrapper[4836]: E0122 10:09:54.534270 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:09:54 crc kubenswrapper[4836]: I0122 10:09:54.684009 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:54 crc kubenswrapper[4836]: I0122 10:09:54.684526 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:54 crc kubenswrapper[4836]: I0122 10:09:54.755452 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.473706 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" exitCode=0 Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.473820 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a"} Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.474823 4836 scope.go:117] "RemoveContainer" containerID="71ddb74a42457a7fdfb4f01738039179de2276d0d842e0346e0a6b83d8de1c62" Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.475804 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:09:55 crc kubenswrapper[4836]: E0122 10:09:55.476273 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.572936 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:55 crc kubenswrapper[4836]: I0122 10:09:55.627297 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:09:57 crc kubenswrapper[4836]: I0122 10:09:57.497205 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qj4w9" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="registry-server" containerID="cri-o://1a15a86442c5722afd98083a83ec9bd60dad67fdc7499f77b2cf5dbe09f06ed2" gracePeriod=2 Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.505636 4836 generic.go:334] "Generic (PLEG): container finished" podID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerID="1a15a86442c5722afd98083a83ec9bd60dad67fdc7499f77b2cf5dbe09f06ed2" exitCode=0 Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.505836 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerDied","Data":"1a15a86442c5722afd98083a83ec9bd60dad67fdc7499f77b2cf5dbe09f06ed2"} Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.506184 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qj4w9" event={"ID":"05df9d3d-2d2b-4531-b352-8d433fc55f51","Type":"ContainerDied","Data":"918ac1f98be0c741969453031eb4ab9d5a3d38e048e91bd0361a209d7866ea09"} Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.506206 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="918ac1f98be0c741969453031eb4ab9d5a3d38e048e91bd0361a209d7866ea09" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.541104 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.616609 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgh2w\" (UniqueName: \"kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w\") pod \"05df9d3d-2d2b-4531-b352-8d433fc55f51\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.616760 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content\") pod \"05df9d3d-2d2b-4531-b352-8d433fc55f51\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.616868 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities\") pod \"05df9d3d-2d2b-4531-b352-8d433fc55f51\" (UID: \"05df9d3d-2d2b-4531-b352-8d433fc55f51\") " Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.618156 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities" (OuterVolumeSpecName: "utilities") pod "05df9d3d-2d2b-4531-b352-8d433fc55f51" (UID: "05df9d3d-2d2b-4531-b352-8d433fc55f51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.623105 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w" (OuterVolumeSpecName: "kube-api-access-mgh2w") pod "05df9d3d-2d2b-4531-b352-8d433fc55f51" (UID: "05df9d3d-2d2b-4531-b352-8d433fc55f51"). InnerVolumeSpecName "kube-api-access-mgh2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.669349 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05df9d3d-2d2b-4531-b352-8d433fc55f51" (UID: "05df9d3d-2d2b-4531-b352-8d433fc55f51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.718633 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgh2w\" (UniqueName: \"kubernetes.io/projected/05df9d3d-2d2b-4531-b352-8d433fc55f51-kube-api-access-mgh2w\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.718665 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:58 crc kubenswrapper[4836]: I0122 10:09:58.718675 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05df9d3d-2d2b-4531-b352-8d433fc55f51-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:09:59 crc kubenswrapper[4836]: I0122 10:09:59.514226 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qj4w9" Jan 22 10:09:59 crc kubenswrapper[4836]: I0122 10:09:59.558684 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:09:59 crc kubenswrapper[4836]: I0122 10:09:59.575498 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qj4w9"] Jan 22 10:10:01 crc kubenswrapper[4836]: I0122 10:10:01.006639 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" path="/var/lib/kubelet/pods/05df9d3d-2d2b-4531-b352-8d433fc55f51/volumes" Jan 22 10:10:07 crc kubenswrapper[4836]: I0122 10:10:07.026411 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:10:07 crc kubenswrapper[4836]: E0122 10:10:07.030751 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:10:17 crc kubenswrapper[4836]: I0122 10:10:17.998328 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:10:17 crc kubenswrapper[4836]: E0122 10:10:17.999177 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:10:29 crc kubenswrapper[4836]: I0122 10:10:29.997566 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:10:29 crc kubenswrapper[4836]: E0122 10:10:29.998397 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:10:44 crc kubenswrapper[4836]: I0122 10:10:44.997095 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:10:44 crc kubenswrapper[4836]: E0122 10:10:44.997711 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:10:55 crc kubenswrapper[4836]: I0122 10:10:55.996783 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:10:55 crc kubenswrapper[4836]: E0122 10:10:55.997769 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:11:09 crc kubenswrapper[4836]: I0122 10:11:09.002968 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:11:09 crc kubenswrapper[4836]: E0122 10:11:09.003976 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:11:20 crc kubenswrapper[4836]: I0122 10:11:20.997607 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:11:20 crc kubenswrapper[4836]: E0122 10:11:20.998604 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:11:31 crc kubenswrapper[4836]: I0122 10:11:31.997027 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:11:31 crc kubenswrapper[4836]: E0122 10:11:31.998036 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:11:47 crc kubenswrapper[4836]: I0122 10:11:47.004838 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:11:47 crc kubenswrapper[4836]: E0122 10:11:47.005768 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:12:01 crc kubenswrapper[4836]: I0122 10:12:01.997474 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:12:01 crc kubenswrapper[4836]: E0122 10:12:01.998180 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:12:13 crc kubenswrapper[4836]: I0122 10:12:13.998191 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:12:14 crc kubenswrapper[4836]: E0122 10:12:13.999530 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:12:28 crc kubenswrapper[4836]: I0122 10:12:28.998449 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:12:29 crc kubenswrapper[4836]: E0122 10:12:28.999386 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:12:41 crc kubenswrapper[4836]: I0122 10:12:41.005754 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:12:41 crc kubenswrapper[4836]: E0122 10:12:41.008100 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:12:53 crc kubenswrapper[4836]: I0122 10:12:53.996783 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:12:53 crc kubenswrapper[4836]: E0122 10:12:53.997643 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:13:07 crc kubenswrapper[4836]: I0122 10:13:07.997308 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:13:07 crc kubenswrapper[4836]: E0122 10:13:07.998120 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:13:18 crc kubenswrapper[4836]: I0122 10:13:18.997448 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:13:18 crc kubenswrapper[4836]: E0122 10:13:18.998142 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:13:32 crc kubenswrapper[4836]: I0122 10:13:32.997884 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:13:32 crc kubenswrapper[4836]: E0122 10:13:32.998489 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:13:47 crc kubenswrapper[4836]: I0122 10:13:47.997570 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:13:48 crc kubenswrapper[4836]: E0122 10:13:47.999582 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:13:58 crc kubenswrapper[4836]: I0122 10:13:58.997836 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:13:59 crc kubenswrapper[4836]: E0122 10:13:58.998649 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:14:12 crc kubenswrapper[4836]: I0122 10:14:12.997689 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:14:12 crc kubenswrapper[4836]: E0122 10:14:12.998591 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:14:23 crc kubenswrapper[4836]: I0122 10:14:23.997825 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:14:23 crc kubenswrapper[4836]: E0122 10:14:23.998764 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:14:34 crc kubenswrapper[4836]: I0122 10:14:34.997091 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:14:34 crc kubenswrapper[4836]: E0122 10:14:34.998013 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:14:45 crc kubenswrapper[4836]: I0122 10:14:45.998177 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:14:45 crc kubenswrapper[4836]: E0122 10:14:45.999099 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.155774 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99"] Jan 22 10:15:00 crc kubenswrapper[4836]: E0122 10:15:00.156991 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="extract-content" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.157009 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="extract-content" Jan 22 10:15:00 crc kubenswrapper[4836]: E0122 10:15:00.157029 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="extract-utilities" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.157037 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="extract-utilities" Jan 22 10:15:00 crc kubenswrapper[4836]: E0122 10:15:00.157077 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="registry-server" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.157086 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="registry-server" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.157263 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="05df9d3d-2d2b-4531-b352-8d433fc55f51" containerName="registry-server" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.157799 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.161468 4836 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.163730 4836 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.183800 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99"] Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.284175 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl578\" (UniqueName: \"kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.284248 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.284359 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.385855 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl578\" (UniqueName: \"kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.386202 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.386346 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.388279 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.394388 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.409436 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl578\" (UniqueName: \"kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578\") pod \"collect-profiles-29484615-85t99\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.478537 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.973919 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99"] Jan 22 10:15:00 crc kubenswrapper[4836]: I0122 10:15:00.997240 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:15:01 crc kubenswrapper[4836]: I0122 10:15:01.225722 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361"} Jan 22 10:15:01 crc kubenswrapper[4836]: I0122 10:15:01.229313 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" event={"ID":"a244a681-9cbc-44e3-a17a-683c4e3a3d3a","Type":"ContainerStarted","Data":"4aa085dd2c47f72d54c3a305386d96eab89a3c6a397b95b1e6441f8165a26be5"} Jan 22 10:15:01 crc kubenswrapper[4836]: I0122 10:15:01.229351 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" event={"ID":"a244a681-9cbc-44e3-a17a-683c4e3a3d3a","Type":"ContainerStarted","Data":"4bca52cf44201287e2f5a6d36345b34ca0c37f31a5ec3b34fd0d5acbdf1fd13b"} Jan 22 10:15:02 crc kubenswrapper[4836]: I0122 10:15:02.241004 4836 generic.go:334] "Generic (PLEG): container finished" podID="a244a681-9cbc-44e3-a17a-683c4e3a3d3a" containerID="4aa085dd2c47f72d54c3a305386d96eab89a3c6a397b95b1e6441f8165a26be5" exitCode=0 Jan 22 10:15:02 crc kubenswrapper[4836]: I0122 10:15:02.241100 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" event={"ID":"a244a681-9cbc-44e3-a17a-683c4e3a3d3a","Type":"ContainerDied","Data":"4aa085dd2c47f72d54c3a305386d96eab89a3c6a397b95b1e6441f8165a26be5"} Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.567230 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.743642 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume\") pod \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.743786 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl578\" (UniqueName: \"kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578\") pod \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.744030 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume\") pod \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\" (UID: \"a244a681-9cbc-44e3-a17a-683c4e3a3d3a\") " Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.744634 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume" (OuterVolumeSpecName: "config-volume") pod "a244a681-9cbc-44e3-a17a-683c4e3a3d3a" (UID: "a244a681-9cbc-44e3-a17a-683c4e3a3d3a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.750532 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578" (OuterVolumeSpecName: "kube-api-access-sl578") pod "a244a681-9cbc-44e3-a17a-683c4e3a3d3a" (UID: "a244a681-9cbc-44e3-a17a-683c4e3a3d3a"). InnerVolumeSpecName "kube-api-access-sl578". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.751223 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a244a681-9cbc-44e3-a17a-683c4e3a3d3a" (UID: "a244a681-9cbc-44e3-a17a-683c4e3a3d3a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.846230 4836 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-config-volume\") on node \"crc\" DevicePath \"\"" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.846261 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl578\" (UniqueName: \"kubernetes.io/projected/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-kube-api-access-sl578\") on node \"crc\" DevicePath \"\"" Jan 22 10:15:03 crc kubenswrapper[4836]: I0122 10:15:03.846272 4836 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a244a681-9cbc-44e3-a17a-683c4e3a3d3a-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 22 10:15:04 crc kubenswrapper[4836]: I0122 10:15:04.258943 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" event={"ID":"a244a681-9cbc-44e3-a17a-683c4e3a3d3a","Type":"ContainerDied","Data":"4bca52cf44201287e2f5a6d36345b34ca0c37f31a5ec3b34fd0d5acbdf1fd13b"} Jan 22 10:15:04 crc kubenswrapper[4836]: I0122 10:15:04.258990 4836 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bca52cf44201287e2f5a6d36345b34ca0c37f31a5ec3b34fd0d5acbdf1fd13b" Jan 22 10:15:04 crc kubenswrapper[4836]: I0122 10:15:04.259029 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-85t99" Jan 22 10:15:04 crc kubenswrapper[4836]: I0122 10:15:04.648282 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r"] Jan 22 10:15:04 crc kubenswrapper[4836]: I0122 10:15:04.658271 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484570-hjs8r"] Jan 22 10:15:05 crc kubenswrapper[4836]: I0122 10:15:05.028017 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faf699f3-dacd-4f97-97b3-0bb300377623" path="/var/lib/kubelet/pods/faf699f3-dacd-4f97-97b3-0bb300377623/volumes" Jan 22 10:16:04 crc kubenswrapper[4836]: I0122 10:16:04.658147 4836 scope.go:117] "RemoveContainer" containerID="1a15a86442c5722afd98083a83ec9bd60dad67fdc7499f77b2cf5dbe09f06ed2" Jan 22 10:16:04 crc kubenswrapper[4836]: I0122 10:16:04.690367 4836 scope.go:117] "RemoveContainer" containerID="f71deaf14fa8797f61d62afa9407b5df72f021238fad41e8dede8d2a5656230e" Jan 22 10:16:04 crc kubenswrapper[4836]: I0122 10:16:04.722020 4836 scope.go:117] "RemoveContainer" containerID="8587bb5ed5bba69f8434b325c4f4f0ef7ed60a0a00673650469a642ab5a88328" Jan 22 10:16:04 crc kubenswrapper[4836]: I0122 10:16:04.754274 4836 scope.go:117] "RemoveContainer" containerID="97bbd6add7cb26a7c2d53a7536546803dff5f26f19b9df6a6d394d97b905c4cb" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.407868 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:32 crc kubenswrapper[4836]: E0122 10:16:32.408737 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a244a681-9cbc-44e3-a17a-683c4e3a3d3a" containerName="collect-profiles" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.408752 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="a244a681-9cbc-44e3-a17a-683c4e3a3d3a" containerName="collect-profiles" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.408909 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="a244a681-9cbc-44e3-a17a-683c4e3a3d3a" containerName="collect-profiles" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.411550 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.433125 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.503762 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.503893 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.503928 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqk8z\" (UniqueName: \"kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.605058 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.605167 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqk8z\" (UniqueName: \"kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.605233 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.605643 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.605660 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.630777 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqk8z\" (UniqueName: \"kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z\") pod \"redhat-operators-nbfsk\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:32 crc kubenswrapper[4836]: I0122 10:16:32.743974 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:33 crc kubenswrapper[4836]: I0122 10:16:33.199638 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:34 crc kubenswrapper[4836]: I0122 10:16:34.061166 4836 generic.go:334] "Generic (PLEG): container finished" podID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerID="5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a" exitCode=0 Jan 22 10:16:34 crc kubenswrapper[4836]: I0122 10:16:34.061207 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerDied","Data":"5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a"} Jan 22 10:16:34 crc kubenswrapper[4836]: I0122 10:16:34.061231 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerStarted","Data":"4ea416a751078a9226b410c0e618f32381232cfb3ff1fe40f2af1c65eccd467e"} Jan 22 10:16:34 crc kubenswrapper[4836]: I0122 10:16:34.063624 4836 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 22 10:16:36 crc kubenswrapper[4836]: I0122 10:16:36.080613 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerStarted","Data":"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd"} Jan 22 10:16:37 crc kubenswrapper[4836]: I0122 10:16:37.095575 4836 generic.go:334] "Generic (PLEG): container finished" podID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerID="e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd" exitCode=0 Jan 22 10:16:37 crc kubenswrapper[4836]: I0122 10:16:37.095631 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerDied","Data":"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd"} Jan 22 10:16:38 crc kubenswrapper[4836]: I0122 10:16:38.102603 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerStarted","Data":"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d"} Jan 22 10:16:38 crc kubenswrapper[4836]: I0122 10:16:38.125695 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nbfsk" podStartSLOduration=2.66057741 podStartE2EDuration="6.125678838s" podCreationTimestamp="2026-01-22 10:16:32 +0000 UTC" firstStartedPulling="2026-01-22 10:16:34.063232133 +0000 UTC m=+2077.423000660" lastFinishedPulling="2026-01-22 10:16:37.528333571 +0000 UTC m=+2080.888102088" observedRunningTime="2026-01-22 10:16:38.122116093 +0000 UTC m=+2081.481884670" watchObservedRunningTime="2026-01-22 10:16:38.125678838 +0000 UTC m=+2081.485447345" Jan 22 10:16:42 crc kubenswrapper[4836]: I0122 10:16:42.744720 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:42 crc kubenswrapper[4836]: I0122 10:16:42.745151 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:43 crc kubenswrapper[4836]: I0122 10:16:43.824746 4836 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nbfsk" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="registry-server" probeResult="failure" output=< Jan 22 10:16:43 crc kubenswrapper[4836]: timeout: failed to connect service ":50051" within 1s Jan 22 10:16:43 crc kubenswrapper[4836]: > Jan 22 10:16:52 crc kubenswrapper[4836]: I0122 10:16:52.820287 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:52 crc kubenswrapper[4836]: I0122 10:16:52.910168 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:53 crc kubenswrapper[4836]: I0122 10:16:53.074218 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.243631 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nbfsk" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="registry-server" containerID="cri-o://c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d" gracePeriod=2 Jan 22 10:16:54 crc kubenswrapper[4836]: E0122 10:16:54.528126 4836 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.688877 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.786400 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities\") pod \"f69802af-115c-4521-aaa8-cf3da2ee339a\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.786463 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqk8z\" (UniqueName: \"kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z\") pod \"f69802af-115c-4521-aaa8-cf3da2ee339a\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.786516 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content\") pod \"f69802af-115c-4521-aaa8-cf3da2ee339a\" (UID: \"f69802af-115c-4521-aaa8-cf3da2ee339a\") " Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.793724 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities" (OuterVolumeSpecName: "utilities") pod "f69802af-115c-4521-aaa8-cf3da2ee339a" (UID: "f69802af-115c-4521-aaa8-cf3da2ee339a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.797262 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z" (OuterVolumeSpecName: "kube-api-access-bqk8z") pod "f69802af-115c-4521-aaa8-cf3da2ee339a" (UID: "f69802af-115c-4521-aaa8-cf3da2ee339a"). InnerVolumeSpecName "kube-api-access-bqk8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.888025 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.888157 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqk8z\" (UniqueName: \"kubernetes.io/projected/f69802af-115c-4521-aaa8-cf3da2ee339a-kube-api-access-bqk8z\") on node \"crc\" DevicePath \"\"" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.952844 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f69802af-115c-4521-aaa8-cf3da2ee339a" (UID: "f69802af-115c-4521-aaa8-cf3da2ee339a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:16:54 crc kubenswrapper[4836]: I0122 10:16:54.989309 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f69802af-115c-4521-aaa8-cf3da2ee339a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.252645 4836 generic.go:334] "Generic (PLEG): container finished" podID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerID="c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d" exitCode=0 Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.252684 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerDied","Data":"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d"} Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.252714 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nbfsk" event={"ID":"f69802af-115c-4521-aaa8-cf3da2ee339a","Type":"ContainerDied","Data":"4ea416a751078a9226b410c0e618f32381232cfb3ff1fe40f2af1c65eccd467e"} Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.252730 4836 scope.go:117] "RemoveContainer" containerID="c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.252758 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nbfsk" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.270884 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.273187 4836 scope.go:117] "RemoveContainer" containerID="e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.278091 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nbfsk"] Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.291677 4836 scope.go:117] "RemoveContainer" containerID="5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.311077 4836 scope.go:117] "RemoveContainer" containerID="c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d" Jan 22 10:16:55 crc kubenswrapper[4836]: E0122 10:16:55.311645 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d\": container with ID starting with c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d not found: ID does not exist" containerID="c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.311683 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d"} err="failed to get container status \"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d\": rpc error: code = NotFound desc = could not find container \"c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d\": container with ID starting with c0f773d20a45f7ee4da12baff314a140d3aa469d281c53a83906403af7e39c3d not found: ID does not exist" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.311710 4836 scope.go:117] "RemoveContainer" containerID="e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd" Jan 22 10:16:55 crc kubenswrapper[4836]: E0122 10:16:55.312283 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd\": container with ID starting with e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd not found: ID does not exist" containerID="e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.312332 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd"} err="failed to get container status \"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd\": rpc error: code = NotFound desc = could not find container \"e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd\": container with ID starting with e1f70709d5a647f6dcf86bd543d42a89ebb526397c373187eaf5424b7b554bcd not found: ID does not exist" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.312368 4836 scope.go:117] "RemoveContainer" containerID="5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a" Jan 22 10:16:55 crc kubenswrapper[4836]: E0122 10:16:55.312684 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a\": container with ID starting with 5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a not found: ID does not exist" containerID="5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a" Jan 22 10:16:55 crc kubenswrapper[4836]: I0122 10:16:55.312717 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a"} err="failed to get container status \"5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a\": rpc error: code = NotFound desc = could not find container \"5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a\": container with ID starting with 5aee822e3611ab42c6db567f17b6810e81c7ff26074d5d3c311085591bcaf40a not found: ID does not exist" Jan 22 10:16:57 crc kubenswrapper[4836]: I0122 10:16:57.007288 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" path="/var/lib/kubelet/pods/f69802af-115c-4521-aaa8-cf3da2ee339a/volumes" Jan 22 10:17:23 crc kubenswrapper[4836]: I0122 10:17:23.897753 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:17:23 crc kubenswrapper[4836]: I0122 10:17:23.898640 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:17:53 crc kubenswrapper[4836]: I0122 10:17:53.898198 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:17:53 crc kubenswrapper[4836]: I0122 10:17:53.899122 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:18:23 crc kubenswrapper[4836]: I0122 10:18:23.897659 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:18:23 crc kubenswrapper[4836]: I0122 10:18:23.898349 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:18:23 crc kubenswrapper[4836]: I0122 10:18:23.898414 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 10:18:23 crc kubenswrapper[4836]: I0122 10:18:23.899245 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 10:18:23 crc kubenswrapper[4836]: I0122 10:18:23.899341 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361" gracePeriod=600 Jan 22 10:18:24 crc kubenswrapper[4836]: I0122 10:18:24.126825 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361" exitCode=0 Jan 22 10:18:24 crc kubenswrapper[4836]: I0122 10:18:24.126862 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361"} Jan 22 10:18:24 crc kubenswrapper[4836]: I0122 10:18:24.127240 4836 scope.go:117] "RemoveContainer" containerID="2e9262c027daa439ce42d1ef8f28f4c06836fd0b21b760014340826a2e3fbd8a" Jan 22 10:18:25 crc kubenswrapper[4836]: I0122 10:18:25.134930 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff"} Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.342397 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:39 crc kubenswrapper[4836]: E0122 10:18:39.343429 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="extract-content" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.343453 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="extract-content" Jan 22 10:18:39 crc kubenswrapper[4836]: E0122 10:18:39.343482 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="registry-server" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.343493 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="registry-server" Jan 22 10:18:39 crc kubenswrapper[4836]: E0122 10:18:39.343514 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="extract-utilities" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.343525 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="extract-utilities" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.343765 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69802af-115c-4521-aaa8-cf3da2ee339a" containerName="registry-server" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.345263 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.364712 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.402640 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.402716 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.402754 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7czg\" (UniqueName: \"kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.503582 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.503632 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.503651 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7czg\" (UniqueName: \"kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.504318 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.504326 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.531013 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7czg\" (UniqueName: \"kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg\") pod \"redhat-marketplace-b9gk9\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:39 crc kubenswrapper[4836]: I0122 10:18:39.672301 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:40 crc kubenswrapper[4836]: I0122 10:18:40.094426 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:40 crc kubenswrapper[4836]: I0122 10:18:40.288204 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerStarted","Data":"42f3bae49144d9ec7b62b160b755d64e641857ecb0047ce7ed59fa1c08340845"} Jan 22 10:18:41 crc kubenswrapper[4836]: I0122 10:18:41.297521 4836 generic.go:334] "Generic (PLEG): container finished" podID="7a7e8b3d-3136-4177-857c-7067d590d795" containerID="61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa" exitCode=0 Jan 22 10:18:41 crc kubenswrapper[4836]: I0122 10:18:41.297628 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerDied","Data":"61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa"} Jan 22 10:18:42 crc kubenswrapper[4836]: I0122 10:18:42.306112 4836 generic.go:334] "Generic (PLEG): container finished" podID="7a7e8b3d-3136-4177-857c-7067d590d795" containerID="76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548" exitCode=0 Jan 22 10:18:42 crc kubenswrapper[4836]: I0122 10:18:42.306187 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerDied","Data":"76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548"} Jan 22 10:18:43 crc kubenswrapper[4836]: I0122 10:18:43.316993 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerStarted","Data":"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8"} Jan 22 10:18:43 crc kubenswrapper[4836]: I0122 10:18:43.335610 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b9gk9" podStartSLOduration=2.881659827 podStartE2EDuration="4.335589485s" podCreationTimestamp="2026-01-22 10:18:39 +0000 UTC" firstStartedPulling="2026-01-22 10:18:41.299640573 +0000 UTC m=+2204.659409100" lastFinishedPulling="2026-01-22 10:18:42.753570231 +0000 UTC m=+2206.113338758" observedRunningTime="2026-01-22 10:18:43.331144916 +0000 UTC m=+2206.690913453" watchObservedRunningTime="2026-01-22 10:18:43.335589485 +0000 UTC m=+2206.695358002" Jan 22 10:18:49 crc kubenswrapper[4836]: I0122 10:18:49.673113 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:49 crc kubenswrapper[4836]: I0122 10:18:49.675758 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:49 crc kubenswrapper[4836]: I0122 10:18:49.732339 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:50 crc kubenswrapper[4836]: I0122 10:18:50.441940 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:50 crc kubenswrapper[4836]: I0122 10:18:50.500919 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:52 crc kubenswrapper[4836]: I0122 10:18:52.377620 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b9gk9" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="registry-server" containerID="cri-o://daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8" gracePeriod=2 Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.341356 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.387875 4836 generic.go:334] "Generic (PLEG): container finished" podID="7a7e8b3d-3136-4177-857c-7067d590d795" containerID="daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8" exitCode=0 Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.387935 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9gk9" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.387950 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerDied","Data":"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8"} Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.388021 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9gk9" event={"ID":"7a7e8b3d-3136-4177-857c-7067d590d795","Type":"ContainerDied","Data":"42f3bae49144d9ec7b62b160b755d64e641857ecb0047ce7ed59fa1c08340845"} Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.388090 4836 scope.go:117] "RemoveContainer" containerID="daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.433329 4836 scope.go:117] "RemoveContainer" containerID="76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.465780 4836 scope.go:117] "RemoveContainer" containerID="61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.492638 4836 scope.go:117] "RemoveContainer" containerID="daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8" Jan 22 10:18:53 crc kubenswrapper[4836]: E0122 10:18:53.493221 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8\": container with ID starting with daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8 not found: ID does not exist" containerID="daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.493254 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8"} err="failed to get container status \"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8\": rpc error: code = NotFound desc = could not find container \"daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8\": container with ID starting with daf0b06bf6245e54ec2356b16fd8ce67d4815aae45f12c86da96360da83a0ba8 not found: ID does not exist" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.493275 4836 scope.go:117] "RemoveContainer" containerID="76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548" Jan 22 10:18:53 crc kubenswrapper[4836]: E0122 10:18:53.493691 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548\": container with ID starting with 76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548 not found: ID does not exist" containerID="76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.493728 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548"} err="failed to get container status \"76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548\": rpc error: code = NotFound desc = could not find container \"76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548\": container with ID starting with 76508541cc0a53a56c50e2bbcbfa41fa90f7b4990d0793c19b1f8f9b65b9c548 not found: ID does not exist" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.493756 4836 scope.go:117] "RemoveContainer" containerID="61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa" Jan 22 10:18:53 crc kubenswrapper[4836]: E0122 10:18:53.494125 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa\": container with ID starting with 61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa not found: ID does not exist" containerID="61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.494158 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa"} err="failed to get container status \"61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa\": rpc error: code = NotFound desc = could not find container \"61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa\": container with ID starting with 61d638d5129c26663252344842a59d5727b0cf2b5449d18e438dc5e5b431c8aa not found: ID does not exist" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.530846 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content\") pod \"7a7e8b3d-3136-4177-857c-7067d590d795\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.530958 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities\") pod \"7a7e8b3d-3136-4177-857c-7067d590d795\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.531002 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7czg\" (UniqueName: \"kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg\") pod \"7a7e8b3d-3136-4177-857c-7067d590d795\" (UID: \"7a7e8b3d-3136-4177-857c-7067d590d795\") " Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.532753 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities" (OuterVolumeSpecName: "utilities") pod "7a7e8b3d-3136-4177-857c-7067d590d795" (UID: "7a7e8b3d-3136-4177-857c-7067d590d795"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.538799 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg" (OuterVolumeSpecName: "kube-api-access-f7czg") pod "7a7e8b3d-3136-4177-857c-7067d590d795" (UID: "7a7e8b3d-3136-4177-857c-7067d590d795"). InnerVolumeSpecName "kube-api-access-f7czg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.550388 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a7e8b3d-3136-4177-857c-7067d590d795" (UID: "7a7e8b3d-3136-4177-857c-7067d590d795"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.632212 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.632242 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7czg\" (UniqueName: \"kubernetes.io/projected/7a7e8b3d-3136-4177-857c-7067d590d795-kube-api-access-f7czg\") on node \"crc\" DevicePath \"\"" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.632252 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a7e8b3d-3136-4177-857c-7067d590d795-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.743828 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:53 crc kubenswrapper[4836]: I0122 10:18:53.755254 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9gk9"] Jan 22 10:18:55 crc kubenswrapper[4836]: I0122 10:18:55.010028 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" path="/var/lib/kubelet/pods/7a7e8b3d-3136-4177-857c-7067d590d795/volumes" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.867740 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:14 crc kubenswrapper[4836]: E0122 10:19:14.868830 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="extract-utilities" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.868853 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="extract-utilities" Jan 22 10:19:14 crc kubenswrapper[4836]: E0122 10:19:14.868878 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="extract-content" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.868889 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="extract-content" Jan 22 10:19:14 crc kubenswrapper[4836]: E0122 10:19:14.868914 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="registry-server" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.868924 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="registry-server" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.869873 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a7e8b3d-3136-4177-857c-7067d590d795" containerName="registry-server" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.871270 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.883641 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.972446 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.972490 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:14 crc kubenswrapper[4836]: I0122 10:19:14.972521 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5ls9\" (UniqueName: \"kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.074198 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5ls9\" (UniqueName: \"kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.074438 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.074469 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.074979 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.076350 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.098778 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5ls9\" (UniqueName: \"kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9\") pod \"community-operators-cjrdt\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.256767 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:15 crc kubenswrapper[4836]: I0122 10:19:15.793172 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:15 crc kubenswrapper[4836]: W0122 10:19:15.812411 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47b416dd_9541_4b62_9a59_597c8b6793ee.slice/crio-fb9be571994386583e32c8114bc95051979272f9a940f6d7481554cd90486218 WatchSource:0}: Error finding container fb9be571994386583e32c8114bc95051979272f9a940f6d7481554cd90486218: Status 404 returned error can't find the container with id fb9be571994386583e32c8114bc95051979272f9a940f6d7481554cd90486218 Jan 22 10:19:16 crc kubenswrapper[4836]: I0122 10:19:16.579862 4836 generic.go:334] "Generic (PLEG): container finished" podID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerID="0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735" exitCode=0 Jan 22 10:19:16 crc kubenswrapper[4836]: I0122 10:19:16.580161 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerDied","Data":"0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735"} Jan 22 10:19:16 crc kubenswrapper[4836]: I0122 10:19:16.580214 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerStarted","Data":"fb9be571994386583e32c8114bc95051979272f9a940f6d7481554cd90486218"} Jan 22 10:19:18 crc kubenswrapper[4836]: I0122 10:19:18.601916 4836 generic.go:334] "Generic (PLEG): container finished" podID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerID="0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe" exitCode=0 Jan 22 10:19:18 crc kubenswrapper[4836]: I0122 10:19:18.604014 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerDied","Data":"0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe"} Jan 22 10:19:19 crc kubenswrapper[4836]: I0122 10:19:19.615710 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerStarted","Data":"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c"} Jan 22 10:19:19 crc kubenswrapper[4836]: I0122 10:19:19.655473 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cjrdt" podStartSLOduration=3.211125169 podStartE2EDuration="5.655458076s" podCreationTimestamp="2026-01-22 10:19:14 +0000 UTC" firstStartedPulling="2026-01-22 10:19:16.582210791 +0000 UTC m=+2239.941979348" lastFinishedPulling="2026-01-22 10:19:19.026543698 +0000 UTC m=+2242.386312255" observedRunningTime="2026-01-22 10:19:19.650854423 +0000 UTC m=+2243.010622940" watchObservedRunningTime="2026-01-22 10:19:19.655458076 +0000 UTC m=+2243.015226593" Jan 22 10:19:25 crc kubenswrapper[4836]: I0122 10:19:25.257682 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:25 crc kubenswrapper[4836]: I0122 10:19:25.258389 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:25 crc kubenswrapper[4836]: I0122 10:19:25.325660 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:25 crc kubenswrapper[4836]: I0122 10:19:25.736284 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:25 crc kubenswrapper[4836]: I0122 10:19:25.791850 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:27 crc kubenswrapper[4836]: I0122 10:19:27.694259 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cjrdt" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="registry-server" containerID="cri-o://6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c" gracePeriod=2 Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.170370 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.310910 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities\") pod \"47b416dd-9541-4b62-9a59-597c8b6793ee\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.311025 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content\") pod \"47b416dd-9541-4b62-9a59-597c8b6793ee\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.311107 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5ls9\" (UniqueName: \"kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9\") pod \"47b416dd-9541-4b62-9a59-597c8b6793ee\" (UID: \"47b416dd-9541-4b62-9a59-597c8b6793ee\") " Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.316353 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9" (OuterVolumeSpecName: "kube-api-access-q5ls9") pod "47b416dd-9541-4b62-9a59-597c8b6793ee" (UID: "47b416dd-9541-4b62-9a59-597c8b6793ee"). InnerVolumeSpecName "kube-api-access-q5ls9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.333611 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities" (OuterVolumeSpecName: "utilities") pod "47b416dd-9541-4b62-9a59-597c8b6793ee" (UID: "47b416dd-9541-4b62-9a59-597c8b6793ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.372416 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47b416dd-9541-4b62-9a59-597c8b6793ee" (UID: "47b416dd-9541-4b62-9a59-597c8b6793ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.412892 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.412937 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47b416dd-9541-4b62-9a59-597c8b6793ee-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.412958 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5ls9\" (UniqueName: \"kubernetes.io/projected/47b416dd-9541-4b62-9a59-597c8b6793ee-kube-api-access-q5ls9\") on node \"crc\" DevicePath \"\"" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.706956 4836 generic.go:334] "Generic (PLEG): container finished" podID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerID="6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c" exitCode=0 Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.707015 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerDied","Data":"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c"} Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.707093 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjrdt" event={"ID":"47b416dd-9541-4b62-9a59-597c8b6793ee","Type":"ContainerDied","Data":"fb9be571994386583e32c8114bc95051979272f9a940f6d7481554cd90486218"} Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.707090 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjrdt" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.707117 4836 scope.go:117] "RemoveContainer" containerID="6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.743451 4836 scope.go:117] "RemoveContainer" containerID="0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.767262 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.771288 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cjrdt"] Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.785215 4836 scope.go:117] "RemoveContainer" containerID="0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.817700 4836 scope.go:117] "RemoveContainer" containerID="6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c" Jan 22 10:19:28 crc kubenswrapper[4836]: E0122 10:19:28.818240 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c\": container with ID starting with 6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c not found: ID does not exist" containerID="6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.818315 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c"} err="failed to get container status \"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c\": rpc error: code = NotFound desc = could not find container \"6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c\": container with ID starting with 6a11f46215c9abb4e5e6779b4ad702b26a8a1ee5e94cfbb4df2338a1888c7c1c not found: ID does not exist" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.818349 4836 scope.go:117] "RemoveContainer" containerID="0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe" Jan 22 10:19:28 crc kubenswrapper[4836]: E0122 10:19:28.819021 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe\": container with ID starting with 0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe not found: ID does not exist" containerID="0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.819078 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe"} err="failed to get container status \"0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe\": rpc error: code = NotFound desc = could not find container \"0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe\": container with ID starting with 0d39483181998a8b36215ab4e7d115dc8f65a5c5ff00ad5de643132e32986cbe not found: ID does not exist" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.819101 4836 scope.go:117] "RemoveContainer" containerID="0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735" Jan 22 10:19:28 crc kubenswrapper[4836]: E0122 10:19:28.819464 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735\": container with ID starting with 0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735 not found: ID does not exist" containerID="0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735" Jan 22 10:19:28 crc kubenswrapper[4836]: I0122 10:19:28.819510 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735"} err="failed to get container status \"0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735\": rpc error: code = NotFound desc = could not find container \"0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735\": container with ID starting with 0fadb57f380619b0e3424fbdc068819ff07ab004116698a7c6a2c920b6772735 not found: ID does not exist" Jan 22 10:19:29 crc kubenswrapper[4836]: I0122 10:19:29.006588 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" path="/var/lib/kubelet/pods/47b416dd-9541-4b62-9a59-597c8b6793ee/volumes" Jan 22 10:20:53 crc kubenswrapper[4836]: I0122 10:20:53.898206 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:20:53 crc kubenswrapper[4836]: I0122 10:20:53.900260 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.564708 4836 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:20:56 crc kubenswrapper[4836]: E0122 10:20:56.567160 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="extract-content" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.567210 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="extract-content" Jan 22 10:20:56 crc kubenswrapper[4836]: E0122 10:20:56.567269 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="registry-server" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.567286 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="registry-server" Jan 22 10:20:56 crc kubenswrapper[4836]: E0122 10:20:56.567304 4836 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="extract-utilities" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.567317 4836 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="extract-utilities" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.567649 4836 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b416dd-9541-4b62-9a59-597c8b6793ee" containerName="registry-server" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.569475 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.601924 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.690910 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.691023 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgw4c\" (UniqueName: \"kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.691196 4836 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.792161 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgw4c\" (UniqueName: \"kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.792246 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.792334 4836 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.792865 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.792935 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.816800 4836 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgw4c\" (UniqueName: \"kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c\") pod \"certified-operators-fkqhk\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:56 crc kubenswrapper[4836]: I0122 10:20:56.893775 4836 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:20:57 crc kubenswrapper[4836]: I0122 10:20:57.180551 4836 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:20:57 crc kubenswrapper[4836]: W0122 10:20:57.185576 4836 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11cd1772_b621_4834_948e_2a7515ad2e71.slice/crio-d8577f62a237e4c95b485c9f750be9d05280d5a7634893d114f48c9ec73b61e9 WatchSource:0}: Error finding container d8577f62a237e4c95b485c9f750be9d05280d5a7634893d114f48c9ec73b61e9: Status 404 returned error can't find the container with id d8577f62a237e4c95b485c9f750be9d05280d5a7634893d114f48c9ec73b61e9 Jan 22 10:20:57 crc kubenswrapper[4836]: I0122 10:20:57.487983 4836 generic.go:334] "Generic (PLEG): container finished" podID="11cd1772-b621-4834-948e-2a7515ad2e71" containerID="683f2e62320b9ec963b1051c87fcb2cb0ccb2d36ddda5d58e1b672f83bf03ff7" exitCode=0 Jan 22 10:20:57 crc kubenswrapper[4836]: I0122 10:20:57.488316 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerDied","Data":"683f2e62320b9ec963b1051c87fcb2cb0ccb2d36ddda5d58e1b672f83bf03ff7"} Jan 22 10:20:57 crc kubenswrapper[4836]: I0122 10:20:57.488345 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerStarted","Data":"d8577f62a237e4c95b485c9f750be9d05280d5a7634893d114f48c9ec73b61e9"} Jan 22 10:20:58 crc kubenswrapper[4836]: I0122 10:20:58.498507 4836 generic.go:334] "Generic (PLEG): container finished" podID="11cd1772-b621-4834-948e-2a7515ad2e71" containerID="c5f69721a9eb75b1623b0301cddd8b5f9f429b4efe0835abffb15c14a82fcc53" exitCode=0 Jan 22 10:20:58 crc kubenswrapper[4836]: I0122 10:20:58.498576 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerDied","Data":"c5f69721a9eb75b1623b0301cddd8b5f9f429b4efe0835abffb15c14a82fcc53"} Jan 22 10:20:59 crc kubenswrapper[4836]: I0122 10:20:59.508290 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerStarted","Data":"cbb0df02456d9de9cae0161823fcb70a47fa26b8e89beb001903eac07d362c53"} Jan 22 10:20:59 crc kubenswrapper[4836]: I0122 10:20:59.536986 4836 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fkqhk" podStartSLOduration=1.947997466 podStartE2EDuration="3.536967051s" podCreationTimestamp="2026-01-22 10:20:56 +0000 UTC" firstStartedPulling="2026-01-22 10:20:57.489657351 +0000 UTC m=+2340.849425878" lastFinishedPulling="2026-01-22 10:20:59.078626946 +0000 UTC m=+2342.438395463" observedRunningTime="2026-01-22 10:20:59.532691227 +0000 UTC m=+2342.892459744" watchObservedRunningTime="2026-01-22 10:20:59.536967051 +0000 UTC m=+2342.896735568" Jan 22 10:21:06 crc kubenswrapper[4836]: I0122 10:21:06.894901 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:06 crc kubenswrapper[4836]: I0122 10:21:06.895616 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:06 crc kubenswrapper[4836]: I0122 10:21:06.966607 4836 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:07 crc kubenswrapper[4836]: I0122 10:21:07.640475 4836 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:07 crc kubenswrapper[4836]: I0122 10:21:07.704844 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:21:09 crc kubenswrapper[4836]: I0122 10:21:09.588696 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fkqhk" podUID="11cd1772-b621-4834-948e-2a7515ad2e71" containerName="registry-server" containerID="cri-o://cbb0df02456d9de9cae0161823fcb70a47fa26b8e89beb001903eac07d362c53" gracePeriod=2 Jan 22 10:21:10 crc kubenswrapper[4836]: I0122 10:21:10.600429 4836 generic.go:334] "Generic (PLEG): container finished" podID="11cd1772-b621-4834-948e-2a7515ad2e71" containerID="cbb0df02456d9de9cae0161823fcb70a47fa26b8e89beb001903eac07d362c53" exitCode=0 Jan 22 10:21:10 crc kubenswrapper[4836]: I0122 10:21:10.600597 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerDied","Data":"cbb0df02456d9de9cae0161823fcb70a47fa26b8e89beb001903eac07d362c53"} Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.187835 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.326695 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content\") pod \"11cd1772-b621-4834-948e-2a7515ad2e71\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.326748 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities\") pod \"11cd1772-b621-4834-948e-2a7515ad2e71\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.326863 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgw4c\" (UniqueName: \"kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c\") pod \"11cd1772-b621-4834-948e-2a7515ad2e71\" (UID: \"11cd1772-b621-4834-948e-2a7515ad2e71\") " Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.327813 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities" (OuterVolumeSpecName: "utilities") pod "11cd1772-b621-4834-948e-2a7515ad2e71" (UID: "11cd1772-b621-4834-948e-2a7515ad2e71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.333124 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c" (OuterVolumeSpecName: "kube-api-access-lgw4c") pod "11cd1772-b621-4834-948e-2a7515ad2e71" (UID: "11cd1772-b621-4834-948e-2a7515ad2e71"). InnerVolumeSpecName "kube-api-access-lgw4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.403787 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11cd1772-b621-4834-948e-2a7515ad2e71" (UID: "11cd1772-b621-4834-948e-2a7515ad2e71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.428752 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgw4c\" (UniqueName: \"kubernetes.io/projected/11cd1772-b621-4834-948e-2a7515ad2e71-kube-api-access-lgw4c\") on node \"crc\" DevicePath \"\"" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.428787 4836 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.428800 4836 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11cd1772-b621-4834-948e-2a7515ad2e71-utilities\") on node \"crc\" DevicePath \"\"" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.614732 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkqhk" event={"ID":"11cd1772-b621-4834-948e-2a7515ad2e71","Type":"ContainerDied","Data":"d8577f62a237e4c95b485c9f750be9d05280d5a7634893d114f48c9ec73b61e9"} Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.614878 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkqhk" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.615100 4836 scope.go:117] "RemoveContainer" containerID="cbb0df02456d9de9cae0161823fcb70a47fa26b8e89beb001903eac07d362c53" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.639590 4836 scope.go:117] "RemoveContainer" containerID="c5f69721a9eb75b1623b0301cddd8b5f9f429b4efe0835abffb15c14a82fcc53" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.943681 4836 scope.go:117] "RemoveContainer" containerID="683f2e62320b9ec963b1051c87fcb2cb0ccb2d36ddda5d58e1b672f83bf03ff7" Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.964241 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:21:11 crc kubenswrapper[4836]: I0122 10:21:11.970378 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fkqhk"] Jan 22 10:21:13 crc kubenswrapper[4836]: I0122 10:21:13.031600 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11cd1772-b621-4834-948e-2a7515ad2e71" path="/var/lib/kubelet/pods/11cd1772-b621-4834-948e-2a7515ad2e71/volumes" Jan 22 10:21:23 crc kubenswrapper[4836]: I0122 10:21:23.897219 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:21:23 crc kubenswrapper[4836]: I0122 10:21:23.897880 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:21:53 crc kubenswrapper[4836]: I0122 10:21:53.897864 4836 patch_prober.go:28] interesting pod/machine-config-daemon-n8vsx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 22 10:21:53 crc kubenswrapper[4836]: I0122 10:21:53.898497 4836 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 22 10:21:53 crc kubenswrapper[4836]: I0122 10:21:53.898572 4836 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" Jan 22 10:21:53 crc kubenswrapper[4836]: I0122 10:21:53.899538 4836 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff"} pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 22 10:21:53 crc kubenswrapper[4836]: I0122 10:21:53.899635 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerName="machine-config-daemon" containerID="cri-o://dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" gracePeriod=600 Jan 22 10:21:54 crc kubenswrapper[4836]: E0122 10:21:54.029459 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:21:55 crc kubenswrapper[4836]: I0122 10:21:54.999407 4836 generic.go:334] "Generic (PLEG): container finished" podID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" exitCode=0 Jan 22 10:21:55 crc kubenswrapper[4836]: I0122 10:21:55.011385 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerDied","Data":"dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff"} Jan 22 10:21:55 crc kubenswrapper[4836]: I0122 10:21:55.011474 4836 scope.go:117] "RemoveContainer" containerID="79b3c93e93e782bf4b90824aabde07777ed3eebf299d2ad6333864efc2bf0361" Jan 22 10:21:55 crc kubenswrapper[4836]: I0122 10:21:55.012205 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:21:55 crc kubenswrapper[4836]: E0122 10:21:55.012647 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:22:05 crc kubenswrapper[4836]: I0122 10:22:05.997443 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:22:05 crc kubenswrapper[4836]: E0122 10:22:05.998165 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:22:18 crc kubenswrapper[4836]: I0122 10:22:18.997254 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:22:18 crc kubenswrapper[4836]: E0122 10:22:18.998540 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:22:32 crc kubenswrapper[4836]: I0122 10:22:32.997938 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:22:33 crc kubenswrapper[4836]: E0122 10:22:32.998900 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:22:45 crc kubenswrapper[4836]: I0122 10:22:44.999780 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:22:45 crc kubenswrapper[4836]: E0122 10:22:45.000818 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:22:58 crc kubenswrapper[4836]: I0122 10:22:58.997347 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:22:58 crc kubenswrapper[4836]: E0122 10:22:58.998271 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:23:12 crc kubenswrapper[4836]: I0122 10:23:12.998696 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:23:13 crc kubenswrapper[4836]: E0122 10:23:12.999625 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:23:27 crc kubenswrapper[4836]: I0122 10:23:27.997114 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:23:27 crc kubenswrapper[4836]: E0122 10:23:27.998930 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:23:41 crc kubenswrapper[4836]: I0122 10:23:41.997666 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:23:42 crc kubenswrapper[4836]: E0122 10:23:41.998818 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:23:52 crc kubenswrapper[4836]: I0122 10:23:52.997440 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:23:52 crc kubenswrapper[4836]: E0122 10:23:52.998334 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:24:07 crc kubenswrapper[4836]: I0122 10:24:07.002704 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:24:07 crc kubenswrapper[4836]: E0122 10:24:07.004648 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:24:20 crc kubenswrapper[4836]: I0122 10:24:20.997289 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:24:20 crc kubenswrapper[4836]: E0122 10:24:20.998398 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:24:35 crc kubenswrapper[4836]: I0122 10:24:35.996944 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:24:35 crc kubenswrapper[4836]: E0122 10:24:35.998599 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:24:47 crc kubenswrapper[4836]: I0122 10:24:47.004526 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:24:47 crc kubenswrapper[4836]: E0122 10:24:47.005349 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:24:58 crc kubenswrapper[4836]: I0122 10:24:58.997750 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:24:59 crc kubenswrapper[4836]: E0122 10:24:58.998726 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:25:11 crc kubenswrapper[4836]: I0122 10:25:11.998529 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:25:12 crc kubenswrapper[4836]: E0122 10:25:11.999649 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:25:25 crc kubenswrapper[4836]: I0122 10:25:25.996382 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:25:25 crc kubenswrapper[4836]: E0122 10:25:25.996894 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:25:40 crc kubenswrapper[4836]: I0122 10:25:40.997505 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:25:40 crc kubenswrapper[4836]: E0122 10:25:40.998429 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:25:53 crc kubenswrapper[4836]: I0122 10:25:53.997815 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:25:54 crc kubenswrapper[4836]: E0122 10:25:53.999017 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:26:01 crc kubenswrapper[4836]: I0122 10:26:01.177966 4836 generic.go:334] "Generic (PLEG): container finished" podID="f0f02dd2-e3a1-4c0e-9b0c-22385660521f" containerID="0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623" exitCode=0 Jan 22 10:26:01 crc kubenswrapper[4836]: I0122 10:26:01.178091 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" event={"ID":"f0f02dd2-e3a1-4c0e-9b0c-22385660521f","Type":"ContainerDied","Data":"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623"} Jan 22 10:26:01 crc kubenswrapper[4836]: I0122 10:26:01.180196 4836 scope.go:117] "RemoveContainer" containerID="0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623" Jan 22 10:26:01 crc kubenswrapper[4836]: I0122 10:26:01.239348 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j5j4k_must-gather-6dcfr_f0f02dd2-e3a1-4c0e-9b0c-22385660521f/gather/0.log" Jan 22 10:26:07 crc kubenswrapper[4836]: I0122 10:26:07.008861 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:26:07 crc kubenswrapper[4836]: E0122 10:26:07.009811 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.441495 4836 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-j5j4k/must-gather-6dcfr"] Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.441812 4836 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" podUID="f0f02dd2-e3a1-4c0e-9b0c-22385660521f" containerName="copy" containerID="cri-o://de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54" gracePeriod=2 Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.447753 4836 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-j5j4k/must-gather-6dcfr"] Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.792292 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j5j4k_must-gather-6dcfr_f0f02dd2-e3a1-4c0e-9b0c-22385660521f/copy/0.log" Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.793456 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.925540 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output\") pod \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.925857 4836 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n675v\" (UniqueName: \"kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v\") pod \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\" (UID: \"f0f02dd2-e3a1-4c0e-9b0c-22385660521f\") " Jan 22 10:26:08 crc kubenswrapper[4836]: I0122 10:26:08.931310 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v" (OuterVolumeSpecName: "kube-api-access-n675v") pod "f0f02dd2-e3a1-4c0e-9b0c-22385660521f" (UID: "f0f02dd2-e3a1-4c0e-9b0c-22385660521f"). InnerVolumeSpecName "kube-api-access-n675v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.020095 4836 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f0f02dd2-e3a1-4c0e-9b0c-22385660521f" (UID: "f0f02dd2-e3a1-4c0e-9b0c-22385660521f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.027267 4836 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.027301 4836 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n675v\" (UniqueName: \"kubernetes.io/projected/f0f02dd2-e3a1-4c0e-9b0c-22385660521f-kube-api-access-n675v\") on node \"crc\" DevicePath \"\"" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.255408 4836 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-j5j4k_must-gather-6dcfr_f0f02dd2-e3a1-4c0e-9b0c-22385660521f/copy/0.log" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.256177 4836 generic.go:334] "Generic (PLEG): container finished" podID="f0f02dd2-e3a1-4c0e-9b0c-22385660521f" containerID="de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54" exitCode=143 Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.256290 4836 scope.go:117] "RemoveContainer" containerID="de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.256382 4836 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-j5j4k/must-gather-6dcfr" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.285282 4836 scope.go:117] "RemoveContainer" containerID="0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.346664 4836 scope.go:117] "RemoveContainer" containerID="de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54" Jan 22 10:26:09 crc kubenswrapper[4836]: E0122 10:26:09.347348 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54\": container with ID starting with de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54 not found: ID does not exist" containerID="de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.347413 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54"} err="failed to get container status \"de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54\": rpc error: code = NotFound desc = could not find container \"de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54\": container with ID starting with de86d840b47b136b9a56a4719460575f9d09dc8088af1d58d1332945b5626f54 not found: ID does not exist" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.347467 4836 scope.go:117] "RemoveContainer" containerID="0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623" Jan 22 10:26:09 crc kubenswrapper[4836]: E0122 10:26:09.348074 4836 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623\": container with ID starting with 0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623 not found: ID does not exist" containerID="0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623" Jan 22 10:26:09 crc kubenswrapper[4836]: I0122 10:26:09.348116 4836 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623"} err="failed to get container status \"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623\": rpc error: code = NotFound desc = could not find container \"0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623\": container with ID starting with 0b2635b30b1cc95758a5bea433a07efe31d9f279b270814a5ff662ca6d5d4623 not found: ID does not exist" Jan 22 10:26:11 crc kubenswrapper[4836]: I0122 10:26:11.008573 4836 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0f02dd2-e3a1-4c0e-9b0c-22385660521f" path="/var/lib/kubelet/pods/f0f02dd2-e3a1-4c0e-9b0c-22385660521f/volumes" Jan 22 10:26:19 crc kubenswrapper[4836]: I0122 10:26:19.996720 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:26:20 crc kubenswrapper[4836]: E0122 10:26:19.997753 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:26:30 crc kubenswrapper[4836]: I0122 10:26:30.997112 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:26:30 crc kubenswrapper[4836]: E0122 10:26:30.998624 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:26:48 crc kubenswrapper[4836]: I0122 10:26:45.997439 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:26:48 crc kubenswrapper[4836]: E0122 10:26:45.998094 4836 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-n8vsx_openshift-machine-config-operator(17ea94f0-dd58-4f5b-b61c-b59c074af1f3)\"" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" podUID="17ea94f0-dd58-4f5b-b61c-b59c074af1f3" Jan 22 10:27:00 crc kubenswrapper[4836]: I0122 10:27:00.996880 4836 scope.go:117] "RemoveContainer" containerID="dc144ea2e5551ef4ae7c3d77961b5381356e9b87806a0f79b6e94a50a88634ff" Jan 22 10:27:01 crc kubenswrapper[4836]: I0122 10:27:01.837786 4836 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-n8vsx" event={"ID":"17ea94f0-dd58-4f5b-b61c-b59c074af1f3","Type":"ContainerStarted","Data":"f65e3c399a266e7dd1df8ba8c5b6403ef93a671a327a15645bdac78f58f09d59"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134375734024461 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134375735017377 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134370011016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134370011015450 5ustar corecore